Subspace Adversarial Training
About
Single-step adversarial training (AT) has received wide attention as it proved to be both efficient and robust. However, a serious problem of catastrophic overfitting exists, i.e., the robust accuracy against projected gradient descent (PGD) attack suddenly drops to 0% during the training. In this paper, we approach this problem from a novel perspective of optimization and firstly reveal the close link between the fast-growing gradient of each sample and overfitting, which can also be applied to understand robust overfitting in multi-step AT. To control the growth of the gradient, we propose a new AT method, Subspace Adversarial Training (Sub-AT), which constrains AT in a carefully extracted subspace. It successfully resolves both kinds of overfitting and significantly boosts the robustness. In subspace, we also allow single-step AT with larger steps and larger radius, further improving the robustness performance. As a result, we achieve state-of-the-art single-step AT performance. Without any regularization term, our single-step AT can reach over 51% robust accuracy against strong PGD-50 attack of radius 8/255 on CIFAR-10, reaching a competitive performance against standard multi-step PGD-10 AT with huge computational advantages. The code is released at https://github.com/nblt/Sub-AT.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-10 (test) | -- | 273 | |
| Adversarial Robustness | CIFAR-10 (test) | -- | 76 | |
| Adversarial Robustness | CIFAR-100 (test) | -- | 46 | |
| Adversarial Robustness | CIFAR-100 | Final Auto-Attack Accuracy37.84 | 12 | |
| Adversarial Robustness | Tiny ImageNet (test) | Natural Accuracy (Best)40.99 | 8 | |
| Adversarial Robustness | CIFAR-10 | -- | 6 | |
| Image Classification | CIFAR-100 (test) | Best Robust Acc41.48 | 4 |