[2602.16256] Color-based Emotion Representation for Speech Emotion Recognition
Summary
This article presents a novel approach to Speech Emotion Recognition (SER) by utilizing color attributes to represent emotions, enhancing the interpretability and performance of SER systems through machine learning techniques.
Why It Matters
The study addresses limitations in traditional SER methods by introducing color-based representations, which could lead to more nuanced and accurate emotion detection in speech. This innovation has implications for various applications, including human-computer interaction, mental health monitoring, and customer service.
Key Takeaways
- Color attributes can effectively represent emotions in speech.
- The study developed regression models for color attributes using machine learning.
- Multitask learning improved the performance of emotion classification tasks.
- Crowdsourced emotional speech corpus enhances the study's validity.
- This approach may lead to more interpretable and diverse emotion recognition systems.
Electrical Engineering and Systems Science > Audio and Speech Processing arXiv:2602.16256 (eess) [Submitted on 18 Feb 2026] Title:Color-based Emotion Representation for Speech Emotion Recognition Authors:Ryotaro Nagase, Ryoichi Takashima, Yoichi Yamashita View a PDF of the paper titled Color-based Emotion Representation for Speech Emotion Recognition, by Ryotaro Nagase and 1 other authors View PDF HTML (experimental) Abstract:Speech emotion recognition (SER) has traditionally relied on categorical or dimensional labels. However, this technique is limited in representing both the diversity and interpretability of emotions. To overcome this limitation, we focus on color attributes, such as hue, saturation, and value, to represent emotions as continuous and interpretable scores. We annotated an emotional speech corpus with color attributes via crowdsourcing and analyzed them. Moreover, we built regression models for color attributes in SER using machine learning and deep learning, and explored the multitask learning of color attribute regression and emotion classification. As a result, we demonstrated the relationship between color attributes and emotions in speech, and successfully developed color attribute regression models for SER. We also showed that multitask learning improved the performance of each task. Comments: Subjects: Audio and Speech Processing (eess.AS); Artificial Intelligence (cs.AI); Sound (cs.SD) Cite as: arXiv:2602.16256 [eess.AS] (or arXiv:2602.16256v1 ...