Skip to content

Speech Emotion Recognition project by using Multi-Layer Perceptron Model with several customized attributes for optimal performance.

Notifications You must be signed in to change notification settings

prernasingh05/CodeClause_Speech_Emotion_Recognition

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 

Repository files navigation

Speech Emotion Recognition

The project aims to develop a classifier that can predict the emotions of a speaker based on an audio file. For this project RAVDESS dataset is used. The RAVDESS contains 24 professional actors (12 female, 12 male), vocalizing two lexically-matched statements in a neutral North American accent. Speech emotions includes calm, happy, sad, angry, fearful, surprise, and disgust expressions. Each expression is produced at two levels of emotional intensity (normal, strong), with an additional neutral expression.

Dataset available at https://www.kaggle.com/datasets/uwrfkaggler/ravdess-emotional-speech-audio

Multi-Layer Perceptron model -

This project is built using a multi-layer Perceptron (MLP) model, which is a type of artificial neural network. The MLP model used in this project consists of several attributes that can be customized for optimal performance. Such as batch_size, alpha, epsilon, hidden_layer_sizes, learning_rate and max_iter.

The waveform of sound files represents the visual depiction of the audio signal over time. It shows how the amplitude of the audio signal changes at each point in time. The waveform provides insights into the characteristics of the audio, such as its intensity, duration, and variations in amplitude.

Some Waveforms of sound files are -

waveform-1

waveform-2

waveform-3

About

Speech Emotion Recognition project by using Multi-Layer Perceptron Model with several customized attributes for optimal performance.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published