Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Transferable Adversarial Attack based on Integrated Gradients

About

The vulnerability of deep neural networks to adversarial examples has drawn tremendous attention from the community. Three approaches, optimizing standard objective functions, exploiting attention maps, and smoothing decision surfaces, are commonly used to craft adversarial examples. By tightly integrating the three approaches, we propose a new and simple algorithm named Transferable Attack based on Integrated Gradients (TAIG) in this paper, which can find highly transferable adversarial examples for black-box attacks. Unlike previous methods using multiple computational terms or combining with other methods, TAIG integrates the three approaches into one single term. Two versions of TAIG that compute their integrated gradients on a straight-line path and a random piecewise linear path are studied. Both versions offer strong transferability and can seamlessly work together with the previous methods. Experimental results demonstrate that TAIG outperforms the state-of-the-art methods. The code will available at https://github.com/yihuang2016/TAIG

Yi Huang, Adams Wai-Kin Kong• 2022

Related benchmarks

TaskDatasetResultRank
Adversarial AttackImageNet (val)--
222
Adversarial AttackImageNet (test)--
101
Adversarial AttackImageNet-1K
Inc-v3ens349
48
Adversarial AttackImageNet 1k 1000 images
Robust Accuracy (Inc-v3)99.7
24
Adversarial AttackImageNet1k 1000 images subset
Model Accuracy (Inception v3)62.4
24
Adversarial AttackImageNet Clean
Success Rate37.6
15
Showing 6 of 6 rows

Other info

Follow for update