Speech Emotion Recognition via Generation using an Attention-based Variational Recurrent Neural Network
Abstract
The last decade has seen an exponential rise in the number of attention-based models for speech emotion recognition (SER). Most of these models use a spectrogram as the input speech representation and the CNN or RNN or convolutional RNN as the key machine learning (ML) component, and learn feature weights to implement attention. We propose an attention-based model for SER that uses MFCC as the input speech representation and a variational RNN (VRNN) as the key ML component. Since the MFCC is of lower dimension than a spectrogram, the model is size- and data-efficient. The VRNN has been used for problems in vision but rarely for SER. Our model is predictive in nature. At each instant, it infers the emotion class and generates the next observation, computes the generation error, and selectively samples (attends to) the locations of high error. Thus, attention emerges in our model, and does not require learning feature weights. This simple model provides interesting insights when evaluated for SER on benchmark datasets. The model can operate on variable length and infinite duration audio files. This work is the first to explore simultaneous generation and recognition for SER, where the generation capability is necessary for efficient recognition.
Publication Title
Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Recommended Citation
Baruah, M., & Banerjee, B. (2022). Speech Emotion Recognition via Generation using an Attention-based Variational Recurrent Neural Network. Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH, 2022-September, 4710-4714. https://doi.org/10.21437/Interspeech.2022-753