Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Efficient and Effective Universal Adversarial Attack against Vision-Language Pre-training Models

About

Vision-language pre-training (VLP) models, trained on large-scale image-text pairs, have become widely used across a variety of downstream vision-and-language (V+L) tasks. This widespread adoption raises concerns about their vulnerability to adversarial attacks. Non-universal adversarial attacks, while effective, are often impractical for real-time online applications due to their high computational demands per data instance. Recently, universal adversarial perturbations (UAPs) have been introduced as a solution, but existing generator-based UAP methods are significantly time-consuming. To overcome the limitation, we propose a direct optimization-based UAP approach, termed DO-UAP, which significantly reduces resource consumption while maintaining high attack performance. Specifically, we explore the necessity of multimodal loss design and introduce a useful data augmentation strategy. Extensive experiments conducted on three benchmark VLP datasets, six popular VLP models, and three classical downstream tasks demonstrate the efficiency and effectiveness of DO-UAP. Specifically, our approach drastically decreases the time consumption by 23-fold while achieving a better attack performance.

Fan Yang, Yihao Huang, Kailong Wang, Ling Shi, Geguang Pu, Yang Liu, Haoyu Wang• 2024

Related benchmarks

TaskDatasetResultRank
Adversarial AttackMantis-Eval
Attack Success Rate64.68
37
Adversarial AttackNLVR2
Attack Success Rate49.68
37
Adversarial AttackBLINK
Attack Success Rate (ASR)72.11
37
Adversarial AttackQ-Bench
Attack Success Rate60.09
37
Adversarial AttackMVBench
ASR63.35
37
Visual Question AnsweringMM-Vet--
27
Visual Question AnsweringLLaVA-Bench
VQA ASR45.6
12
Visual Question AnsweringMantis-Eval
ASR49.23
12
Showing 8 of 8 rows

Other info

Follow for update