Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

If your data distribution shifts, use self-learning

About

We demonstrate that self-learning techniques like entropy minimization and pseudo-labeling are simple and effective at improving performance of a deployed computer vision model under systematic domain shifts. We conduct a wide range of large-scale experiments and show consistent improvements irrespective of the model architecture, the pre-training technique or the type of distribution shift. At the same time, self-learning is simple to use in practice because it does not require knowledge or access to the original training data or scheme, is robust to hyperparameter choices, is straight-forward to implement and requires only a few adaptation epochs. This makes self-learning techniques highly attractive for any practitioner who applies machine learning algorithms in the real world. We present state-of-the-art adaptation results on CIFAR10-C (8.5% error), ImageNet-C (22.0% mCE), ImageNet-R (17.4% error) and ImageNet-A (14.8% error), theoretically study the dynamics of self-supervised adaptation methods and propose a new classification dataset (ImageNet-D) which is challenging even with adaptation.

Evgenia Rusak, Steffen Schneider, George Pachitariu, Luisa Eck, Peter Gehler, Oliver Bringmann, Wieland Brendel, Matthias Bethge• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet (val)--
188
Image ClassificationImageNet-A (test)--
175
Image ClassificationSTL-10--
146
Image ClassificationImageNet-R (test)--
118
Image ClassificationImageNet-C (test)
mCE (Mean Corruption Error)22
116
Image ClassificationCIFAR-10-C (test)--
61
Image ClassificationImageNet-D--
36
Image ClassificationCamelyon17-WILDS out-of-distribution (val)
Accuracy93
16
Image ClassificationRxRx1 ID WILDS (val)
Top-1 Accuracy34.8
16
Image ClassificationRxRx1 OOD WILDS (test)
Top-1 Acc29.4
16
Showing 10 of 22 rows

Other info

Code

Follow for update