Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Tensor Fusion Network for Multimodal Sentiment Analysis

About

Multimodal sentiment analysis is an increasingly popular research area, which extends the conventional language-based definition of sentiment analysis to a multimodal setup where other relevant modalities accompany language. In this paper, we pose the problem of multimodal sentiment analysis as modeling intra-modality and inter-modality dynamics. We introduce a novel model, termed Tensor Fusion Network, which learns both such dynamics end-to-end. The proposed approach is tailored for the volatile nature of spoken language in online videos as well as accompanying gestures and voice. In the experiments, our model outperforms state-of-the-art approaches for both multimodal and unimodal sentiment analysis.

Amir Zadeh, Minghai Chen, Soujanya Poria, Erik Cambria, Louis-Philippe Morency• 2017

Related benchmarks

TaskDatasetResultRank
Multimodal Sentiment AnalysisCMU-MOSI (test)
F180.7
238
Multimodal Sentiment AnalysisCMU-MOSEI (test)
F1 Score79.11
206
Emotion Recognition in ConversationIEMOCAP (test)
Weighted Average F1 Score55.13
154
Emotion Recognition in ConversationMELD
Weighted Avg F157.74
137
Emotion Recognition in ConversationMELD (test)
Weighted F157.74
118
Alzheimer stage classificationADNI
AUC74.24
116
Multimodal Sentiment AnalysisCMU-MOSI standard (test)
Accuracy77.1
62
Multimodal Sentiment AnalysisCMU-MOSI
MAE0.901
59
Multimodal Sentiment AnalysisMOSEI (test)
MAE0.593
49
Survival PredictionMulti-modal Survival Dataset (test)
C-index0.86
44
Showing 10 of 56 rows

Other info

Follow for update