Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Attention Bottlenecks for Multimodal Fusion

About

Humans perceive the world by concurrently processing and fusing high-dimensional inputs from multiple modalities such as vision and audio. Machine perception models, in stark contrast, are typically modality-specific and optimised for unimodal benchmarks, and hence late-stage fusion of final representations or predictions from each modality (`late-fusion') is still a dominant paradigm for multimodal video classification. Instead, we introduce a novel transformer based architecture that uses `fusion bottlenecks' for modality fusion at multiple layers. Compared to traditional pairwise self-attention, our model forces information between different modalities to pass through a small number of bottleneck latents, requiring the model to collate and condense the most relevant information in each modality and only share what is necessary. We find that such a strategy improves fusion performance, at the same time reducing computational cost. We conduct thorough ablation studies, and achieve state-of-the-art results on multiple audio-visual classification benchmarks including Audioset, Epic-Kitchens and VGGSound. All code and models will be released.

Arsha Nagrani, Shan Yang, Anurag Arnab, Aren Jansen, Cordelia Schmid, Chen Sun• 2021

Related benchmarks

TaskDatasetResultRank
Action RecognitionKinetics-400
Top-1 Acc80.8
481
Action RecognitionUCF101 (test)
Accuracy91.8
307
Text-to-Video RetrievalMSR-VTT (test)
R@13.8
255
Audio ClassificationAudioSet 20K
mAP31.3
128
Action RecognitionEPIC-KITCHENS 100 (test)
Top-1 Verb Acc64.8
101
Multimodal Multilabel ClassificationMM-IMDB (test)
Macro F159.6
87
Audio ClassificationVGG-Sound
Top-1 Accuracy52.3
83
Audio ClassificationAudioSet 2M
mAP44.3
79
ClassificationAudioSet (test)
mAP44.3
57
Audio ClassificationAudioSet
mAP41.5
54
Showing 10 of 53 rows

Other info

Follow for update