Make Some Noise: Reliable and Efficient Single-Step Adversarial Training
About
Recently, Wong et al. showed that adversarial training with single-step FGSM leads to a characteristic failure mode named Catastrophic Overfitting (CO), in which a model becomes suddenly vulnerable to multi-step attacks. Experimentally they showed that simply adding a random perturbation prior to FGSM (RS-FGSM) could prevent CO. However, Andriushchenko and Flammarion observed that RS-FGSM still leads to CO for larger perturbations, and proposed a computationally expensive regularizer (GradAlign) to avoid it. In this work, we methodically revisit the role of noise and clipping in single-step adversarial training. Contrary to previous intuitions, we find that using a stronger noise around the clean sample combined with \textit{not clipping} is highly effective in avoiding CO for large perturbation radii. We then propose Noise-FGSM (N-FGSM) that, while providing the benefits of single-step adversarial training, does not suffer from CO. Empirical analyses on a large suite of experiments show that N-FGSM is able to match or surpass the performance of previous state-of-the-art GradAlign, while achieving 3x speed-up. Code can be found in https://github.com/pdejorge/N-FGSM
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet-100 | -- | 84 | |
| Adversarial Robustness | CIFAR-10 (test) | -- | 76 | |
| Image Classification | CIFAR-10 (test) | Natural Accuracy80.4 | 48 | |
| Image Classification | CIFAR100 (test) | Natural Accuracy54.92 | 40 | |
| Image Classification | CIFAR10 (test) | Natural Accuracy80.48 | 40 | |
| Image Classification | CIFAR100 | Robust Accuracy22.68 | 34 | |
| Image Classification | CIFAR-10 (test) | Accuracy81.21 | 31 | |
| Image Classification | SVHN | Accuracy (Natural)95.09 | 30 | |
| Image Classification | Tiny ImageNet (test) | Standard Accuracy44.96 | 22 | |
| Image Classification | CIFAR-100 (test) | SA55.4 | 22 |