Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Domain-Agnostic Mutual Prompting for Unsupervised Domain Adaptation

About

Conventional Unsupervised Domain Adaptation (UDA) strives to minimize distribution discrepancy between domains, which neglects to harness rich semantics from data and struggles to handle complex domain shifts. A promising technique is to leverage the knowledge of large-scale pre-trained vision-language models for more guided adaptation. Despite some endeavors, current methods often learn textual prompts to embed domain semantics for source and target domains separately and perform classification within each domain, limiting cross-domain knowledge transfer. Moreover, prompting only the language branch lacks flexibility to adapt both modalities dynamically. To bridge this gap, we propose Domain-Agnostic Mutual Prompting (DAMP) to exploit domain-invariant semantics by mutually aligning visual and textual embeddings. Specifically, the image contextual information is utilized to prompt the language branch in a domain-agnostic and instance-conditioned way. Meanwhile, visual prompts are imposed based on the domain-agnostic textual prompt to elicit domain-invariant visual embeddings. These two branches of prompts are learned mutually with a cross-attention module and regularized with a semantic-consistency loss and an instance-discrimination contrastive loss. Experiments on three UDA benchmarks demonstrate the superiority of DAMP over state-of-the-art approaches.

Zhekai Du, Xinyao Li, Fengling Li, Ke Lu, Lei Zhu, Jingjing Li• 2024

Related benchmarks

TaskDatasetResultRank
Unsupervised Domain AdaptationOffice-Home
Average Accuracy78.2
238
Domain GeneralizationPACS (test)
Average Accuracy97.4
225
Image ClassificationDomainNet
Accuracy (ClipArt)69.7
161
Image ClassificationOffice-Home
Average Accuracy87.1
142
Domain GeneralizationOffice-Home (test)
Average Accuracy85
106
Domain GeneralizationVLCS (test)--
62
Image ClassificationOffice-Home
Average Accuracy79.2
59
Image ClassificationVisDA (val)
Plane Accuracy97.3
44
Domain GeneralizationTerraIncognita (test)
Accuracy53.7
40
Closed-set Source-Free Domain AdaptationVisDA Sy→Re
Accuracy (Sy→Re)88.4
37
Showing 10 of 16 rows

Other info

Follow for update