If your data distribution shifts, use self-learning
About
We demonstrate that self-learning techniques like entropy minimization and pseudo-labeling are simple and effective at improving performance of a deployed computer vision model under systematic domain shifts. We conduct a wide range of large-scale experiments and show consistent improvements irrespective of the model architecture, the pre-training technique or the type of distribution shift. At the same time, self-learning is simple to use in practice because it does not require knowledge or access to the original training data or scheme, is robust to hyperparameter choices, is straight-forward to implement and requires only a few adaptation epochs. This makes self-learning techniques highly attractive for any practitioner who applies machine learning algorithms in the real world. We present state-of-the-art adaptation results on CIFAR10-C (8.5% error), ImageNet-C (22.0% mCE), ImageNet-R (17.4% error) and ImageNet-A (14.8% error), theoretically study the dynamics of self-supervised adaptation methods and propose a new classification dataset (ImageNet-D) which is challenging even with adaptation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet (val) | -- | 188 | |
| Image Classification | ImageNet-A (test) | -- | 154 | |
| Image Classification | STL-10 | -- | 128 | |
| Image Classification | ImageNet-C (test) | mCE (Mean Corruption Error)22 | 110 | |
| Image Classification | ImageNet-R (test) | -- | 105 | |
| Image Classification | CIFAR-10-C (test) | -- | 61 | |
| Image Classification | CCC Hard 1.0 (test) | Mean Accuracy0.67 | 12 | |
| Image Classification | CCC Medium 1.0 (test) | Mean Accuracy2.7 | 12 | |
| Image Classification | CCC Easy 1.0 (test) | Accuracy7.5 | 12 | |
| Image Classification | Camelyon17-WILDS (val) | Accuracy97.6 | 11 |