DINE: Domain Adaptation from Single and Multiple Black-box Predictors
About
To ease the burden of labeling, unsupervised domain adaptation (UDA) aims to transfer knowledge in previous and related labeled datasets (sources) to a new unlabeled dataset (target). Despite impressive progress, prior methods always need to access the raw source data and develop data-dependent alignment approaches to recognize the target samples in a transductive learning manner, which may raise privacy concerns from source individuals. Several recent studies resort to an alternative solution by exploiting the well-trained white-box model from the source domain, yet, it may still leak the raw data through generative adversarial learning. This paper studies a practical and interesting setting for UDA, where only black-box source models (i.e., only network predictions are available) are provided during adaptation in the target domain. To solve this problem, we propose a new two-step knowledge adaptation framework called DIstill and fine-tuNE (DINE). Taking into consideration the target data structure, DINE first distills the knowledge from the source predictor to a customized target model, then fine-tunes the distilled model to further fit the target domain. Besides, neural networks are not required to be identical across domains in DINE, even allowing effective adaptation on a low-resource device. Empirical results on three UDA scenarios (i.e., single-source, multi-source, and partial-set) confirm that DINE achieves highly competitive performance compared to state-of-the-art data-dependent approaches. Code is available at \url{https://github.com/tim-learn/DINE/}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unsupervised Domain Adaptation | ImageCLEF-DA | Average Accuracy91.9 | 104 | |
| Image Classification | CIFAR-10-C (test) | Accuracy (Clean)73.86 | 61 | |
| Image Classification | CIFAR100-C (test) | Robustness Accuracy40.52 | 29 | |
| object recognition | Office 59 (test) | Acc (A->D)95.5 | 21 | |
| object recognition | VisDA-C | Accuracy (Plane)96.6 | 21 | |
| Partial-set Unsupervised Domain Adaptation | Office-Home Partial-set 76 | Accuracy (Ar -> Cl)67.8 | 20 | |
| Multi-source closed-set UDA | Office-Home target domains Ar, Cl, Pr, Re | Accuracy (Ar)83.6 | 16 | |
| Image Classification | OOD-CV 0% occlusion 1.0 | Top-1 Accuracy (Combined)83.5 | 15 | |
| Multi-source closed-set UDA | Office target domains A, D, W | Acc (Target A)82.4 | 13 | |
| Multi-source closed-set UDA | Office-Caltech target domains A, C, D, W | Acc (Target A)0.968 | 13 |