Skip to content

Omicron02/Emotion-Detector-Using-Voice-as-an-Input

Repository files navigation

Emotion Detector Using Voice as an Input

The aim of the project is to analyze the user's voice and determine their mood which can then be used for various purposes.

Companies can provide a more personalised experience by recommending products based on the person's mood. One such example being in voice assistants. When the user asks it to play a song for example, it can play a song based on the user's mood.

Datasets:

RAVDESS: This dataset includes around 1400 audio file input from 24 different actors. 12 male and 12 female where these actors record short audios in 8 different emotions i.e 1 = neutral, 2 = calm, 3 = happy, 4 = sad, 5 = angry, 6 = fearful, 7 = disgust, 8 = surprised. Each audio file is named in such a way that the 7th character is consistent with the different emotions that they represent.

Audio files:

Tested out the audio files by plotting out the waveform and a spectrogram to see the sample audio files.
Waveform


Spectrogram

Predictions

These are a sample of the actual vs predicted values.


Testing out with live voices

In order to test out our model on voices that were completely different than what we have in our training and test data, we recorded our own voices with dfferent emotions and predicted the outcomes.
Waveform of recorded audio.

Output of user's voice data.

Contributors

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors