Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Unlabeled Data Improves Adversarial Robustness

About

We demonstrate, theoretically and empirically, that adversarial robustness can significantly benefit from semisupervised learning. Theoretically, we revisit the simple Gaussian model of Schmidt et al. that shows a sample complexity gap between standard and robust classification. We prove that unlabeled data bridges this gap: a simple semisupervised learning procedure (self-training) achieves high robust accuracy using the same number of labels required for achieving high standard accuracy. Empirically, we augment CIFAR-10 with 500K unlabeled images sourced from 80 Million Tiny Images and use robust self-training to outperform state-of-the-art robust accuracies by over 5 points in (i) $\ell_\infty$ robustness against several strong attacks via adversarial training and (ii) certified $\ell_2$ and $\ell_\infty$ robustness via randomized smoothing. On SVHN, adding the dataset's own extra training set with the labels removed provides gains of 4 to 10 points, within 1 point of the gain from using the extra labels.

Yair Carmon, Aditi Raghunathan, Ludwig Schmidt, Percy Liang, John C. Duchi• 2019

Related benchmarks

TaskDatasetResultRank
Adversarial RobustnessCIFAR-10 (test)--
76
Image ClassificationCIFAR-10
AA Accuracy59.53
38
Image ClassificationSVHN WRN-16-8 (test)
Accuracy (Clean)97.4
28
Image ClassificationCIFAR-10 WRN-28-10 (test)
Clean Accuracy89.7
28
Image ClassificationCIFAR-10 512-image subset (test)
Clean Accuracy89.67
26
Image ClassificationCIFAR100 (test)
Natural Accuracy47.54
16
Image ClassificationCIFAR-10 (test)
AutoAttack Accuracy59.53
14
Robust Image ClassificationRobustBench (test)
RA59
12
Image ClassificationCIFAR-10 (test)
Natural Accuracy73.22
9
Image ClassificationImageNet 32-100
Accuracy (Natural)28.64
8
Showing 10 of 12 rows

Other info

Follow for update