Multi-Task Learning with Auxiliary Speaker Identification for Conversational Emotion Recognition
About
Conversational emotion recognition (CER) has attracted increasing interests in the natural language processing (NLP) community. Different from the vanilla emotion recognition, effective speaker-sensitive utterance representation is one major challenge for CER. In this paper, we exploit speaker identification (SI) as an auxiliary task to enhance the utterance representation in conversations. By this method, we can learn better speaker-aware contextual representations from the additional SI corpus. Experiments on two benchmark datasets demonstrate that the proposed architecture is highly effective for CER, obtaining new state-of-the-art results on two datasets.
Jingye Li, Meishan Zhang, Donghong Ji, Yijiang Liu• 2020
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Emotion Recognition in Conversation | MELD | Weighted Avg F161.9 | 137 | |
| Emotion Detection | EmoryNLP (test) | -- | 96 | |
| Emotion Recognition | MELD (test) | W-Avg F1 (7-cls)61.9 | 26 | |
| Conversational Emotion Recognition | MELD (test) | Macro F1 Score61.9 | 12 |
Showing 4 of 4 rows