Cluster Alignment with a Teacher for Unsupervised Domain Adaptation
About
Deep learning methods have shown promise in unsupervised domain adaptation, which aims to leverage a labeled source domain to learn a classifier for the unlabeled target domain with a different distribution. However, such methods typically learn a domain-invariant representation space to match the marginal distributions of the source and target domains, while ignoring their fine-level structures. In this paper, we propose Cluster Alignment with a Teacher (CAT) for unsupervised domain adaptation, which can effectively incorporate the discriminative clustering structures in both domains for better adaptation. Technically, CAT leverages an implicit ensembling teacher model to reliably discover the class-conditional structure in the feature space for the unlabeled target domain. Then CAT forces the features of both the source and the target domains to form discriminative class-conditional clusters and aligns the corresponding clusters across domains. Empirical results demonstrate that CAT achieves state-of-the-art results in several unsupervised domain adaptation scenarios.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Office-31 | Average Accuracy87.6 | 261 | |
| Domain Adaptation | Office-31 unsupervised adaptation standard | Accuracy (A to W)94.4 | 162 | |
| Unsupervised Domain Adaptation | ImageCLEF-DA | Average Accuracy87.3 | 104 | |
| Domain Adaptation | Image-CLEF DA (test) | Average Accuracy87.3 | 76 | |
| object recognition | Office (standard) | Accuracy (A to W)94.4 | 55 | |
| Domain Adaptation | SVHN to MNIST (test) | Accuracy98.8 | 53 | |
| Domain Adaptation Classification | Office-31 (test) | A -> W Accuracy91.1 | 31 | |
| Domain Adaptation | Office31 standard (test) | Standard Accuracy (A->D)90.8 | 28 | |
| Digit Recognition | Digits SVHN, MNIST, USPS (test) | Accuracy (S->M)98.8 | 26 | |
| Image Classification | Digits | Average Accuracy89.9 | 23 |