MACER: Attack-free and Scalable Robust Training via Maximizing Certified Radius
About
Adversarial training is one of the most popular ways to learn robust models but is usually attack-dependent and time costly. In this paper, we propose the MACER algorithm, which learns robust models without using adversarial training but performs better than all existing provable l2-defenses. Recent work shows that randomized smoothing can be used to provide a certified l2 radius to smoothed classifiers, and our algorithm trains provably robust smoothed classifiers via MAximizing the CErtified Radius (MACER). The attack-free characteristic makes MACER faster to train and easier to optimize. In our experiments, we show that our method can be applied to modern deep neural networks on a wide range of datasets, including Cifar-10, ImageNet, MNIST, and SVHN. For all tasks, MACER spends less training time than state-of-the-art adversarial training algorithms, and the learned models achieve larger average certified radius.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | MNIST | -- | 263 | |
| Certified Image Classification | MNIST (test) | Certified Accuracy (r=0.00)99.33 | 27 | |
| Image Classification Certified Robustness | MNIST (test) | Overall ACR1.598 | 27 | |
| Certified Robustness | CIFAR-10 (test) | -- | 26 | |
| Certified Robust Classification | CIFAR-10 official (test) | ACR0.691 | 14 |