Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multimodal Prompting with Missing Modalities for Visual Recognition

About

In this paper, we tackle two challenges in multimodal learning for visual recognition: 1) when missing-modality occurs either during training or testing in real-world situations; and 2) when the computation resources are not available to finetune on heavy transformer models. To this end, we propose to utilize prompt learning and mitigate the above two challenges together. Specifically, our modality-missing-aware prompts can be plugged into multimodal transformers to handle general missing-modality cases, while only requiring less than 1% learnable parameters compared to training the entire model. We further explore the effect of different prompt configurations and analyze the robustness to missing modality. Extensive experiments are conducted to show the effectiveness of our prompt learning framework that improves the performance under various missing-modality cases, while alleviating the requirement of heavy model re-training. Code is available.

Yi-Lun Lee, Yi-Hsuan Tsai, Wei-Chen Chiu, Chen-Yu Lee• 2023

Related benchmarks

TaskDatasetResultRank
Multimodal Sentiment AnalysisCMU-MOSI (test)
F163.7
238
Multimodal Sentiment AnalysisCMU-MOSEI (test)
F1 Score68.7
206
Image ClassificationFood101 (test)
Accuracy86.72
87
Multimodal Multilabel ClassificationMM-IMDB (test)
Macro F150.52
87
Emotion RecognitionIEMOCAP
Accuracy65.77
71
Hateful Meme DetectionHateful Memes (test)
AUROC0.6607
67
Multimodal Sentiment AnalysisMOSI
Accuracy80.74
54
Disease DiagnosisCXR 20 (test)
F1-Macro71.23
36
Disease DiagnosisODIR 21 (test)
F1 Macro98.95
36
Multimodal ClassificationMST Missing Modalities
Accuracy100
28
Showing 10 of 34 rows

Other info

Code

Follow for update