Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

EmotionIC: emotional inertia and contagion-driven dependency modeling for emotion recognition in conversation

About

Emotion Recognition in Conversation (ERC) has attracted growing attention in recent years as a result of the advancement and implementation of human-computer interface technologies. In this paper, we propose an emotional inertia and contagion-driven dependency modeling approach (EmotionIC) for ERC task. Our EmotionIC consists of three main components, i.e., Identity Masked Multi-Head Attention (IMMHA), Dialogue-based Gated Recurrent Unit (DiaGRU), and Skip-chain Conditional Random Field (SkipCRF). Compared to previous ERC models, EmotionIC can model a conversation more thoroughly at both the feature-extraction and classification levels. The proposed model attempts to integrate the advantages of attention- and recurrence-based methods at the feature-extraction level. Specifically, IMMHA is applied to capture identity-based global contextual dependencies, while DiaGRU is utilized to extract speaker- and temporal-aware local contextual information. At the classification level, SkipCRF can explicitly mine complex emotional flows from higher-order neighboring utterances in the conversation. Experimental results show that our method can significantly outperform the state-of-the-art models on four benchmark datasets. The ablation studies confirm that our modules can effectively model emotional inertia and contagion.

Yingjian Liu, Jiang Li, Xiaoping Wang, Zhigang Zeng• 2023

Related benchmarks

TaskDatasetResultRank
Emotion Recognition in ConversationIEMOCAP (test)
Weighted Average F1 Score69.61
168
Emotion Recognition in ConversationMELD
Weighted Avg F166.4
137
Conversational Emotion RecognitionIEMOCAP
Weighted Average F1 Score69.5
129
Emotion RecognitionIEMOCAP
Accuracy69.44
115
Emotion DetectionEmoryNLP (test)
Weighted-F10.3965
96
Dialogue Emotion DetectionEmoryNLP
Weighted Avg F140.01
80
Emotion Recognition in ConversationMELD standard (test)
Weighted F166.4
19
Cross-scenario Multimodal Emotion Recognition in ConversationsIEMOCAP -> MELD noise rate 40% (test)
Joy Accuracy37.53
15
Cross-scenario Multimodal Emotion RecognitionMELD -> IEMOCAP 20% Noise (test)
Joy Accuracy33.12
15
Cross-scenario Multimodal Emotion Recognition in ConversationsMELD -> IEMOCAP noise rate 40% (test)
Joy Accuracy24.74
15
Showing 10 of 13 rows

Other info

Follow for update