Partial Transfer Learning with Selective Adversarial Networks
About
Adversarial learning has been successfully embedded into deep networks to learn transferable features, which reduce distribution discrepancy between the source and target domains. Existing domain adversarial networks assume fully shared label space across domains. In the presence of big data, there is strong motivation of transferring both classification and representation models from existing big domains to unknown small domains. This paper introduces partial transfer learning, which relaxes the shared label space assumption to that the target label space is only a subspace of the source label space. Previous methods typically match the whole source domain to the target domain, which are prone to negative transfer for the partial transfer problem. We present Selective Adversarial Network (SAN), which simultaneously circumvents negative transfer by selecting out the outlier source classes and promotes positive transfer by maximally matching the data distributions in the shared label space. Experiments demonstrate that our models exceed state-of-the-art results for partial transfer learning tasks on several benchmark datasets.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Office-Home (test) | Mean Accuracy65.3 | 199 | |
| Domain Adaptation | Office-Home | Average Accuracy65.3 | 111 | |
| Partial Domain Adaptation | Office-Home | Average Accuracy65.3 | 97 | |
| Image Classification | Office-31 (test) | Avg Accuracy95 | 93 | |
| Partial Domain Adaptation | Office-31 | Avg Accuracy94.96 | 52 | |
| Partial Domain Adaptation | ImageNet-Caltech | Average Accuracy77.8 | 32 | |
| Partial Domain Adaptation | Office-31 (test) | Accuracy (A -> W)93.9 | 25 | |
| Partial Domain Adaptation | VisDA Synthetic-12 -> Real-6 2017 | Accuracy69.7 | 25 | |
| Image Classification | Office-Home 65->25 | Ar->Cl Accuracy44.4 | 16 | |
| Image Classification | VisDA-C 12->6 | Accuracy (R->S)69.7 | 11 |