Learning Language-guided Adaptive Hyper-modality Representation for Multimodal Sentiment Analysis
About
Though Multimodal Sentiment Analysis (MSA) proves effective by utilizing rich information from multiple sources (e.g., language, video, and audio), the potential sentiment-irrelevant and conflicting information across modalities may hinder the performance from being further improved. To alleviate this, we present Adaptive Language-guided Multimodal Transformer (ALMT), which incorporates an Adaptive Hyper-modality Learning (AHL) module to learn an irrelevance/conflict-suppressing representation from visual and audio features under the guidance of language features at different scales. With the obtained hyper-modality representation, the model can obtain a complementary and joint representation through multimodal fusion for effective MSA. In practice, ALMT achieves state-of-the-art performance on several popular datasets (e.g., MOSI, MOSEI and CH-SIMS) and an abundance of ablation demonstrates the validity and necessity of our irrelevance/conflict suppression mechanism.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multimodal Sentiment Analysis | CMU-MOSI | MAE0.683 | 59 | |
| Multimodal Sentiment Analysis | MOSEI (test) | MAE0.526 | 49 | |
| Multimodal Sentiment Analysis | MOSI (test) | MAE0.683 | 34 | |
| Multimodal Sentiment Analysis | SIMS (test) | MAE0.5912 | 22 | |
| Multimodal Sentiment Analysis | CH-SIMS | F1 Score77.6 | 18 | |
| Multimodal Sentiment Analysis | MOSI | F1 Score85.1 | 12 | |
| Multimodal Sentiment Analysis | CH-SIMS (test) | Acc (2-class)81.19 | 8 |