Self-ensembling for visual domain adaptation
About
This paper explores the use of self-ensembling for visual domain adaptation problems. Our technique is derived from the mean teacher variant (Tarvainen et al., 2017) of temporal ensembling (Laine et al;, 2017), a technique that achieved state of the art results in the area of semi-supervised learning. We introduce a number of modifications to their approach for challenging domain adaptation scenarios and evaluate its effectiveness. Our approach achieves state of the art results in a variety of benchmarks, including our winning entry in the VISDA-2017 visual domain adaptation challenge. In small image benchmarks, our algorithm not only outperforms prior art, but can also achieve accuracy that is close to that of a classifier trained in a supervised fashion.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Domain Adaptation | VisDA 2017 (test) | Mean Class Accuracy92.8 | 98 | |
| Image Classification | VisDA 2017 (test) | Class Accuracy (Plane)95.9 | 83 | |
| Image Classification | Office-10 + Caltech-10 | Average Accuracy89.7 | 77 | |
| Image Classification | SVHN to MNIST (test) | Accuracy98.6 | 66 | |
| Digit Classification | MNIST -> USPS (test) | Accuracy88.14 | 65 | |
| Image Classification | MNIST -> USPS (test) | Accuracy98.26 | 64 | |
| Image Classification | USPS -> MNIST (test) | Accuracy99.54 | 63 | |
| Digit Classification | USPS → MNIST target (test) | Accuracy92.35 | 58 | |
| Domain Adaptation | VisDA 2017 (val) | Mean Accuracy86.6 | 52 | |
| Image Classification | SYN SIGNS to GTSRB (test) | Accuracy99.37 | 49 |