Test-Time Training with Self-Supervision for Generalization under Distribution Shifts
About
In this paper, we propose Test-Time Training, a general approach for improving the performance of predictive models when training and test data come from different distributions. We turn a single unlabeled test sample into a self-supervised learning problem, on which we update the model parameters before making a prediction. This also extends naturally to data in an online stream. Our simple approach leads to improvements on diverse image classification benchmarks aimed at evaluating robustness to distribution shifts.
Yu Sun, Xiaolong Wang, Zhuang Liu, John Miller, Alexei A. Efros, Moritz Hardt• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-10C Severity Level 5 (test) | Average Error Rate (Severity 5)19.1 | 127 | |
| Image Classification | ImageNet-C level 5 | Avg Top-1 Acc (ImageNet-C L5)83.6 | 110 | |
| Robot Manipulation | LIBERO Object | Success Rate97.2 | 70 | |
| Image Classification | CIFAR-100C Level 5 (test) | Gaussian Acc83.8 | 51 | |
| Robotic Manipulation | LIBERO Long | Success Rate89.1 | 44 | |
| Diabetic Retinopathy Classification | DEEPDR (test) | Accuracy0.524 | 30 | |
| Robot Manipulation | LIBERO Average | Success Rate94.7 | 26 | |
| Image Classification | CIFAR-10C level 5 (test) | Mean Error12.1 | 26 | |
| Diabetic Retinopathy (DR) grading | APTOS | Accuracy51.7 | 25 | |
| Diabetic Retinopathy (DR) grading | Messidor | ACC64.8 | 22 |
Showing 10 of 41 rows