Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: Emotion recognition from speech plays a vital role in the development of
empathetic human-computer interaction systems. This paper presents a
comparative analysis of lightweight transformer-based models, DistilHuBERT and
PaSST, by classifying six core emotions from the CREMA-D dataset. We benchmark
their performance against a traditional CNN-LSTM baseline model using MFCC
features. DistilHuBERT demonstrates superior accuracy (70.64%) and F1 score
(70.36%) while maintaining an exceptionally small model size (0.02 MB),
outperforming both PaSST and the baseline. Furthermore, we conducted an
ablation study on three variants of the PaSST, Linear, MLP, and Attentive
Pooling heads, to understand the effect of classification head architecture on
model performance. Our results indicate that PaSST with an MLP head yields the
best performance among its variants but still falls short of DistilHuBERT.
Among the emotion classes, angry is consistently the most accurately detected,
while disgust remains the most challenging. These findings suggest that
lightweight transformers like DistilHuBERT offer a compelling solution for
real-time speech emotion recognition on edge devices. The code is available at:
https://github.com/luckymaduabuchi/Emotion-detection-.
Key Contributions
Compares lightweight transformer models (DistilHuBERT, PaSST) for speech emotion detection, finding DistilHuBERT to be superior in accuracy and F1 score with minimal model size. It also performs an ablation study on PaSST's classification heads.
Business Value
Enables the development of more empathetic and responsive AI systems (e.g., virtual assistants, customer service bots) by accurately detecting user emotions from speech.