Understanding Catastrophic Overfitting in Single-step Adversarial Training
About
Although fast adversarial training has demonstrated both robustness and efficiency, the problem of "catastrophic overfitting" has been observed. This is a phenomenon in which, during single-step adversarial training, the robust accuracy against projected gradient descent (PGD) suddenly decreases to 0% after a few epochs, whereas the robust accuracy against fast gradient sign method (FGSM) increases to 100%. In this paper, we demonstrate that catastrophic overfitting is very closely related to the characteristic of single-step adversarial training which uses only adversarial examples with the maximum perturbation, and not all adversarial examples in the adversarial direction, which leads to decision boundary distortion and a highly curved loss surface. Based on this observation, we propose a simple method that not only prevents catastrophic overfitting, but also overrides the belief that it is difficult to prevent multi-step adversarial attacks with single-step adversarial training.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Adversarial Robustness | CIFAR-10 (test) | -- | 76 | |
| Adversarial Robustness | CIFAR-100 (test) | -- | 46 | |
| Image Classification | CIFAR-10 (test) | Accuracy90.02 | 31 | |
| Image Classification | Tiny ImageNet (test) | Standard Accuracy57.93 | 22 | |
| Image Classification | CIFAR-100 (test) | SA69.79 | 22 | |
| Image Classification | CIFAR-100 WRN34-10 (test) | SA Success Rate60.66 | 22 | |
| Image Classification | CIFAR10 (test) | Accuracy (Natural)90.45 | 21 |