Click4Ai

543.

Hard

Emotion Recognition from Audio

In this problem, you will use a deep learning approach to classify audio clips into different emotions. You will need to use a pre-trained model and fine-tune it on a dataset of labeled audio clips.

**Example:** Given an audio clip, predict the emotion it represents.

**Constraints:** The input audio clip is a 10-second WAV file, and the model should output a probability distribution over 7 emotions: happy, sad, angry, fearful, surprised, disgusted, and neutral.

**Note:** You can use the Librosa library to load and preprocess the audio data, and the PyTorch library to build and train the model.

Test Cases

Test Case 1
Input: audio_clip.wav
Expected: [[0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7]]
Test Case 2
Input: audio_clip2.wav
Expected: [[0.8, 0.1, 0.05, 0.02, 0.01, 0.005, 0.0005]]
+ 3 hidden test cases