Emotion Recognition from Audio
In this problem, you will use a deep learning approach to classify audio clips into different emotions. You will need to use a pre-trained model and fine-tune it on a dataset of labeled audio clips.
**Example:** Given an audio clip, predict the emotion it represents.
**Constraints:** The input audio clip is a 10-second WAV file, and the model should output a probability distribution over 7 emotions: happy, sad, angry, fearful, surprised, disgusted, and neutral.
**Note:** You can use the Librosa library to load and preprocess the audio data, and the PyTorch library to build and train the model.
Test Cases
Test Case 1
Input:
audio_clip.wavExpected:
[[0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7]]Test Case 2
Input:
audio_clip2.wavExpected:
[[0.8, 0.1, 0.05, 0.02, 0.01, 0.005, 0.0005]]+ 3 hidden test cases