Skip to content

ubaidkhan08/Speech-Emotion-Recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Speech-Emotion-Recognition

An emotion classifier built using standard audio data processing and deep learning algorithms. Here, we have 4 different datasets with a 12,000+ audio files and a plethora of voice actors to generalize the model and avoid overfitting over a certain accent. Due to the sheer complexity of SER (Speech Emotion Recognition), the accuracy will be 60-70% only. However. we've tried to give you a brief comparison of various decisions over the accuracy.

Datasets used:

  1. Crowd-sourced Emotional Mutimodal Actors Dataset (Crema-D)
  2. Ryerson Audio-Visual Database of Emotional Speech and Song (Ravdess)
  3. Surrey Audio-Visual Expressed Emotion (Savee)
  4. Toronto emotional speech set (Tess)

Algorithm used: Sequential with 1D convolution layer (Conv1D) & Maxpooling

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published