Researchers have developed a new model called ML-SAN to improve emotion recognition in conversations by accounting for individual differences in expression. This Multi-Level Speaker-Adaptive Network uses a three-stage process to calibrate input features, adapt modality trust based on speaker identity, and maintain speaker consistency in the latent space. Tests on the MELD and IEMOCAP datasets indicate that ML-SAN performs better, particularly with less common sentiment categories and diverse speakers. AI
Summary written by gemini-2.5-flash-lite from 2 sources. How we write summaries →
IMPACT Improves multimodal emotion recognition by adapting to individual speaker expression styles, enhancing machine empathy.
RANK_REASON This is a research paper introducing a novel model for emotion recognition.