Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multimodal Transformer for Unaligned Multimodal Language Sequences

About

Human language is often multimodal, which comprehends a mixture of natural language, facial gestures, and acoustic behaviors. However, two major challenges in modeling such multimodal human language time-series data exist: 1) inherent data non-alignment due to variable sampling rates for the sequences from each modality; and 2) long-range dependencies between elements across modalities. In this paper, we introduce the Multimodal Transformer (MulT) to generically address the above issues in an end-to-end manner without explicitly aligning the data. At the heart of our model is the directional pairwise crossmodal attention, which attends to interactions between multimodal sequences across distinct time steps and latently adapt streams from one modality to another. Comprehensive experiments on both aligned and non-aligned multimodal time-series show that our model outperforms state-of-the-art methods by a large margin. In addition, empirical analysis suggests that correlated crossmodal signals are able to be captured by the proposed crossmodal attention mechanism in MulT.

Yao-Hung Hubert Tsai, Shaojie Bai, Paul Pu Liang, J. Zico Kolter, Louis-Philippe Morency, Ruslan Salakhutdinov• 2019

Related benchmarks

TaskDatasetResultRank
Multimodal Sentiment AnalysisCMU-MOSI (test)
F183.9
238
Multimodal Sentiment AnalysisCMU-MOSEI (test)
F1 Score82.3
206
Emotion Recognition in ConversationIEMOCAP (test)--
154
Alzheimer stage classificationADNI
AUC72.43
116
Emotion RecognitionIEMOCAP--
71
Multimodal Sentiment AnalysisCMU-MOSI
MAE0.846
59
Multimodal Sentiment AnalysisMOSEI (test)
MAE0.58
49
Mortality PredictionMIMIC-IV (test)
AUC67.35
43
Sentiment AnalysisCMU-MOSEI (test)
Acc (2-class)82.5
40
Facial Expression RecognitionAFEW (test)
Accuracy55.87
35
Showing 10 of 83 rows
...

Other info

Code

Follow for update