Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Improve Unsupervised Domain Adaptation with Mixup Training

About

Unsupervised domain adaptation studies the problem of utilizing a relevant source domain with abundant labels to build predictive modeling for an unannotated target domain. Recent work observe that the popular adversarial approach of learning domain-invariant features is insufficient to achieve desirable target domain performance and thus introduce additional training constraints, e.g. cluster assumption. However, these approaches impose the constraints on source and target domains individually, ignoring the important interplay between them. In this work, we propose to enforce training constraints across domains using mixup formulation to directly address the generalization performance for target data. In order to tackle potentially huge domain discrepancy, we further propose a feature-level consistency regularizer to facilitate the inter-domain constraint. When adding intra-domain mixup and domain adversarial learning, our general framework significantly improves state-of-the-art performance on several important tasks from both image classification and human activity recognition.

Shen Yan, Huan Song, Nanxiang Li, Lincan Zou, Liu Ren• 2020

Related benchmarks

TaskDatasetResultRank
Domain GeneralizationVLCS
Accuracy78.7
238
Image ClassificationPACS
Overall Average Accuracy63.9
230
Domain GeneralizationPACS (test)
Average Accuracy64.2
225
Domain GeneralizationPACS
Accuracy (Art)87.4
221
Domain GeneralizationDomainBed (test)
VLCS Accuracy77.7
110
Domain GeneralizationDomainBed v1.0 (test)
Average Accuracy78.1
71
Image ClassificationOfficeHome DomainBed suite (test)
Accuracy61.7
45
Domain GeneralizationDomainNet DomainBed (test)
Clipart Accuracy48.9
37
Image ClassificationDomainBed
PACS Accuracy79.2
33
Domain GeneralizationOfficeHome DomainBed (test)
Accuracy68
29
Showing 10 of 33 rows

Other info

Follow for update