Improving Adversarial Transferability via Neuron Attribution-Based Attacks
About
Deep neural networks (DNNs) are known to be vulnerable to adversarial examples. It is thus imperative to devise effective attack algorithms to identify the deficiencies of DNNs beforehand in security-sensitive applications. To efficiently tackle the black-box setting where the target model's particulars are unknown, feature-level transfer-based attacks propose to contaminate the intermediate feature outputs of local models, and then directly employ the crafted adversarial samples to attack the target model. Due to the transferability of features, feature-level attacks have shown promise in synthesizing more transferable adversarial samples. However, existing feature-level attacks generally employ inaccurate neuron importance estimations, which deteriorates their transferability. To overcome such pitfalls, in this paper, we propose the Neuron Attribution-based Attack (NAA), which conducts feature-level attacks with more accurate neuron importance estimations. Specifically, we first completely attribute a model's output to each neuron in a middle layer. We then derive an approximation scheme of neuron attribution to tremendously reduce the computation overhead. Finally, we weight neurons based on their attribution results and launch feature-level attacks. Extensive experiments confirm the superiority of our approach to the state-of-the-art benchmarks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet ILSVRC 1000 images 2012 (val) | Attack Success Rate98.2 | 180 | |
| Adversarial Attack | ImageNet Adversarially Trained Models | Robust Accuracy (Inc-v3 adv)76 | 20 | |
| Adversarial Attack | ImageNet Undefended Models | Inc-v3 Robustness Score98.8 | 20 | |
| Black-box Adversarial Attack | ImageNet (test) | -- | 13 | |
| Adversarial Attack | ImageNet | Inception v3 Robust Accuracy22.86 | 5 |