Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Certified Adversarial Robustness via Randomized Smoothing

About

We show how to turn any classifier that classifies well under Gaussian noise into a new classifier that is certifiably robust to adversarial perturbations under the $\ell_2$ norm. This "randomized smoothing" technique has been proposed recently in the literature, but existing guarantees are loose. We prove a tight robustness guarantee in $\ell_2$ norm for smoothing with Gaussian noise. We use randomized smoothing to obtain an ImageNet classifier with e.g. a certified top-1 accuracy of 49% under adversarial perturbations with $\ell_2$ norm less than 0.5 (=127/255). No certified defense has been shown feasible on ImageNet except for smoothing. On smaller-scale datasets where competing approaches to certified $\ell_2$ robustness are viable, smoothing delivers higher certified accuracies. Our strong empirical results suggest that randomized smoothing is a promising direction for future research into adversarially robust classification. Code and models are available at http://github.com/locuslab/smoothing.

Jeremy M Cohen, Elan Rosenfeld, J. Zico Kolter• 2019

Related benchmarks

TaskDatasetResultRank
Image ClassificationMNIST--
263
Image ClassificationCIFAR-10 corrupted (test)
Acc88.3
30
Certified Image ClassificationMNIST (test)
Certified Accuracy (r=0.00)99.25
27
Image Classification Certified RobustnessMNIST (test)
Overall ACR1.62
27
Certified RobustnessCIFAR-10 (test)
Accuracy (Standard)92.7
26
Image ClassificationCIFAR-10.1 1.0 (test)
Accuracy76.7
14
Certified Robust ClassificationCIFAR-10 official (test)
ACR0.525
14
Certified AccuracyCIFAR-10 (test)
Certified Accuracy (r=0.0)65
9
Image ClassificationImageNet sub-sampled 500 samples (val)
ACR0.875
8
Image ClassificationImageNet 10-class subset (test)
Certified Accuracy (eps=0.00)93.4
4
Showing 10 of 10 rows

Other info

Follow for update