Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Can audio-visual integration strengthen robustness under multimodal attacks?

About

In this paper, we propose to make a systematic study on machines multisensory perception under attacks. We use the audio-visual event recognition task against multimodal adversarial attacks as a proxy to investigate the robustness of audio-visual learning. We attack audio, visual, and both modalities to explore whether audio-visual integration still strengthens perception and how different fusion mechanisms affect the robustness of audio-visual models. For interpreting the multimodal interactions under attacks, we learn a weakly-supervised sound source visual localization model to localize sounding regions in videos. To mitigate multimodal attacks, we propose an audio-visual defense approach based on an audio-visual dissimilarity constraint and external feature memory banks. Extensive experiments demonstrate that audio-visual models are susceptible to multimodal adversarial attacks; audio-visual integration could decrease the model robustness rather than strengthen under multimodal attacks; even a weakly-supervised sound source visual localization model can be successfully fooled; our defense method can improve the invulnerability of audio-visual networks without significantly sacrificing clean model performance.

Yapeng Tian, Chenliang Xu• 2021

Related benchmarks

TaskDatasetResultRank
Audio-visual event recognitionAVE (test)
AV Accuracy71.39
20
Audio-visual event recognitionMIT-MUSIC (test)
AV Accuracy91.35
14
Audio-visual event recognitionKinetics-Sounds (test)
AV Score72.71
14
Audio-Visual ClassificationAVE
AV Score71.39
6
Showing 4 of 4 rows

Other info

Code

Follow for update