Towards Generalized Certified Robustness with Multi-Norm Training
About
Existing certified training methods can only train models to be robust against a certain perturbation type (e.g. $l_\infty$ or $l_2$). However, an $l_\infty$ certifiably robust model may not be certifiably robust against $l_2$ perturbation (and vice versa) and also has low robustness against other perturbations (e.g. geometric and patch transformation). By constructing a theoretical framework to analyze and mitigate the tradeoff, we propose the first multi-norm certified training framework \textbf{CURE}, consisting of several multi-norm certified training methods, to attain better \emph{union robustness} when training from scratch or fine-tuning a pre-trained certified model. Inspired by our theoretical findings, we devise bound alignment and connect natural training with certified training for better union robustness. Compared with SOTA-certified training, \textbf{CURE} improves union robustness to $32.0\%$ on MNIST, $25.8\%$ on CIFAR-10, and $10.6\%$ on TinyImagenet across different epsilon values. It leads to better generalization on a diverse set of challenging unseen geometric and patch perturbations to $6.8\%$ and $16.0\%$ on CIFAR-10. Overall, our contributions pave a path towards \textit{generalized certified robustness}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Certified Robustness | CIFAR-10 (test) | -- | 26 | |
| Image Classification | MNIST (test) | Test Accuracy99.3 | 24 | |
| Image Classification | CIFAR-10 (test) | Clean Accuracy79.4 | 7 | |
| Image Classification | CIFAR-10 | Clean Accuracy53 | 7 | |
| Image Classification | TinyImageNet | Clean Accuracy30.5 | 7 | |
| Image Classification | MNIST | Clean Accuracy98.7 | 7 | |
| Certified Robustness | CIFAR-100 (test) | Clean Accuracy42.5 | 6 | |
| Certified Robustness | CIFAR-10 | Certified Radius (L-inf)61.2 | 4 | |
| Text Classification | SST-2 PWWS | Robust Accuracy28.4 | 4 | |
| Text Classification | SST-2 TextFooler | Robust Accuracy17.6 | 4 |