Test-Time Training with Masked Autoencoders
About
Test-time training adapts to a new test distribution on the fly by optimizing a model for each test input using self-supervision. In this paper, we use masked autoencoders for this one-sample learning problem. Empirically, our simple method improves generalization on many visual benchmarks for distribution shifts. Theoretically, we characterize this improvement in terms of the bias-variance trade-off.
Yossi Gandelsman, Yu Sun, Xinlei Chen, Alexei A. Efros• 2022
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | ImageNet A | Top-1 Acc21.3 | 553 | |
| Image Classification | ImageNet-R | Top-1 Acc38.9 | 474 | |
| Image Classification | ImageNet-C level 5 | Avg Top-1 Acc (ImageNet-C L5)45.9 | 61 | |
| Image Classification | ImageNet-C 1.0 (test) | Accuracy (Average)45.92 | 53 | |
| Spatio-temporal forecasting | NYC Taxi (test) | MAE14.902 | 41 | |
| Image Classification | ImageNet-C level 3 (test) | Acc (Brightness)75.8 | 34 | |
| Traffic Flow Forecasting | PeMS07 Jan 2017 - Aug 2017 (6:2:2) | MAE20.051 | 12 | |
| Traffic Flow Forecasting | BayArea Jan 2019 - Dec 2019 (5:2:3) | MAE15.409 | 12 | |
| Traffic Flow Forecasting (inflow) | NYCTaxi grid-based (test) | MAE17.221 | 12 | |
| Traffic Flow Forecasting (inflow) | T-Drive grid-based (test) | MAE22.164 | 12 |
Showing 10 of 16 rows