Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SmoothMix: Training Confidence-calibrated Smoothed Classifiers for Certified Robustness

About

Randomized smoothing is currently a state-of-the-art method to construct a certifiably robust classifier from neural networks against $\ell_2$-adversarial perturbations. Under the paradigm, the robustness of a classifier is aligned with the prediction confidence, i.e., the higher confidence from a smoothed classifier implies the better robustness. This motivates us to rethink the fundamental trade-off between accuracy and robustness in terms of calibrating confidences of a smoothed classifier. In this paper, we propose a simple training scheme, coined SmoothMix, to control the robustness of smoothed classifiers via self-mixup: it trains on convex combinations of samples along the direction of adversarial perturbation for each input. The proposed procedure effectively identifies over-confident, near off-class samples as a cause of limited robustness in case of smoothed classifiers, and offers an intuitive way to adaptively set a new decision boundary between these samples for better robustness. Our experimental results demonstrate that the proposed method can significantly improve the certified $\ell_2$-robustness of smoothed classifiers compared to existing state-of-the-art robust training methods.

Jongheon Jeong, Sejun Park, Minkyu Kim, Heung-Chang Lee, Doguk Kim, Jinwoo Shin• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationMNIST--
263
Certified Image ClassificationMNIST (test)
Certified Accuracy (r=0.00)99.45
27
Image Classification Certified RobustnessMNIST (test)
Overall ACR1.823
27
Certified RobustnessCIFAR-10 (test)--
26
Certified Robust ClassificationCIFAR-10 official (test)
ACR0.737
14
Image ClassificationImageNet sub-sampled 500 samples (val)
ACR1.047
8
Showing 6 of 6 rows

Other info

Code

Follow for update