Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

SMIL: Multimodal Learning with Severely Missing Modality

About

A common assumption in multimodal learning is the completeness of training data, i.e., full modalities are available in all training examples. Although there exists research endeavor in developing novel methods to tackle the incompleteness of testing data, e.g., modalities are partially missing in testing examples, few of them can handle incomplete training modalities. The problem becomes even more challenging if considering the case of severely missing, e.g., 90% training examples may have incomplete modalities. For the first time in the literature, this paper formally studies multimodal learning with missing modality in terms of flexibility (missing modalities in training, testing, or both) and efficiency (most training data have incomplete modality). Technically, we propose a new method named SMIL that leverages Bayesian meta-learning in uniformly achieving both objectives. To validate our idea, we conduct a series of experiments on three popular benchmarks: MM-IMDb, CMU-MOSI, and avMNIST. The results prove the state-of-the-art performance of SMIL over existing methods and generative baselines including autoencoders and generative adversarial networks. Our code is available at https://github.com/mengmenm/SMIL.

Mengmeng Ma, Jian Ren, Long Zhao, Sergey Tulyakov, Cathy Wu, Xi Peng• 2021

Related benchmarks

TaskDatasetResultRank
Multimodal Sentiment AnalysisCMU-MOSEI (test)--
332
Multimodal Sentiment AnalysisCMU-MOSI (test)--
316
Survival PredictionTCGA-LUAD
C-index0.695
154
Survival PredictionTCGA-UCEC
C-index0.74
142
Emotion RecognitionIEMOCAP--
115
Multimodal Multilabel ClassificationMM-IMDB (test)--
87
Readmission predictionMIMIC IV
AUC-ROC0.6894
70
Multimodal Sentiment AnalysisMOSEI (test)--
49
Arousal Emotion RecognitionDEAP (test)
Accuracy88.35
47
Survival AnalysisTCGA-GBMLGG
C-index0.844
44
Showing 10 of 22 rows

Other info

Follow for update