Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Vision Transformers are Parameter-Efficient Audio-Visual Learners

About

Vision transformers (ViTs) have achieved impressive results on various computer vision tasks in the last several years. In this work, we study the capability of frozen ViTs, pretrained only on visual data, to generalize to audio-visual data without finetuning any of its original parameters. To do so, we propose a latent audio-visual hybrid (LAVISH) adapter that adapts pretrained ViTs to audio-visual tasks by injecting a small number of trainable parameters into every layer of a frozen ViT. To efficiently fuse visual and audio cues, our LAVISH adapter uses a small set of latent tokens, which form an attention bottleneck, thus, eliminating the quadratic cost of standard cross-attention. Compared to the existing modality-specific audio-visual methods, our approach achieves competitive or even better performance on various audio-visual tasks while using fewer tunable parameters and without relying on costly audio pretraining or external audio encoders. Our code is available at https://genjib.github.io/project_page/LAVISH/

Yan-Bo Lin, Yi-Lin Sung, Jie Lei, Mohit Bansal, Gedas Bertasius• 2022

Related benchmarks

TaskDatasetResultRank
Action RecognitionUCF101 (test)
Accuracy92.6
307
Audio-Visual Question AnsweringMUSIC-AVQA 1.0 (test)
AV Localis Accuracy81.31
96
Audio-Visual Question AnsweringMUSIC-AVQA (test)
Acc (Avg)74.46
59
Audio Question AnsweringMUSIC-AVQA 1.0 (test)
Counting Accuracy82.09
43
Audio-Visual Event LocalizationAVE (test)
Accuracy81.1
37
Audio-Visual Event LocalizationAVE
Accuracy81.1
35
Audio-Visual SegmentationAVSBench MS3 (test)
Jaccard Index (IoU)49.8
30
Audio-Video Question AnsweringMUSIC-AVQA
AV Temporal Acc0.691
19
Audio-Visual Question AnsweringMUSIC-AVQA balanced v2.0 (test)
Total Accuracy73.18
18
Audio-Visual Question AnsweringMUSIC-AVQA Bias v2.0 (test)
Total Accuracy74.59
18
Showing 10 of 18 rows

Other info

Code

Follow for update