Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Understanding Zero-Shot Adversarial Robustness for Large-Scale Models

About

Pretrained large-scale vision-language models like CLIP have exhibited strong generalization over unseen tasks. Yet imperceptible adversarial perturbations can significantly reduce CLIP's performance on new tasks. In this work, we identify and explore the problem of \emph{adapting large-scale models for zero-shot adversarial robustness}. We first identify two key factors during model adaption -- training losses and adaptation methods -- that affect the model's zero-shot adversarial robustness. We then propose a text-guided contrastive adversarial training loss, which aligns the text embeddings and the adversarial visual features with contrastive learning on a small set of training data. We apply this training loss to two adaption methods, model finetuning and visual prompt tuning. We find that visual prompt tuning is more effective in the absence of texts, while finetuning wins in the existence of text guidance. Overall, our approach significantly improves the zero-shot adversarial robustness over CLIP, seeing an average improvement of over 31 points over ImageNet and 15 zero-shot datasets. We hope this work can shed light on understanding the zero-shot adversarial robustness of large-scale models.

Chengzhi Mao, Scott Geng, Junfeng Yang, Xin Wang, Carl Vondrick• 2022

Related benchmarks

TaskDatasetResultRank
Visual Question AnsweringVQA v2--
1362
Image ClassificationImageNet A
Top-1 Acc8.5
654
Image ClassificationImageNet V2
Top-1 Acc52.3
611
Image ClassificationEuroSAT
Accuracy1.88
569
Image ClassificationFlowers102
Accuracy15.62
558
Image ClassificationDTD
Accuracy8.62
485
Image ClassificationFood101
Accuracy6.98
457
Image ClassificationUCF101
Top-1 Acc9.29
455
Image ClassificationSUN397
Accuracy53.29
441
Image ClassificationStanfordCars
Accuracy6.64
312
Showing 10 of 255 rows
...

Other info

Follow for update