Abstract: This research categorized a deep learning based framework for multimodal emotion recognition in conversa- tions, while addressing class imbalance in emotion datasets. The framework combined text, audio, and visual modalities with methods of imbalanced learning to help the recognition of minority emotions. Evaluations across benchmark datasets for the multimodal framework achieved improvements over baseline methods overall, and with respect to the underrepresented classes.

Keywords: Multimodal emotion recognition, unbalanced learning, deep learning


Downloads: PDF | DOI: 10.17148/IARJSET.2026.13101

How to Cite:

[1] Sindhu B M, Deepthi M B, Sanika G S, Shrusti, Ramya B Kanoji, "Addressing Data Imbalance in Multimodal Conversational Emotion Analysis," International Advanced Research Journal in Science, Engineering and Technology (IARJSET), DOI: 10.17148/IARJSET.2026.13101

Open chat