Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Domain Adaptation via Bidirectional Cross-Attention Transformer

About

Domain Adaptation (DA) aims to leverage the knowledge learned from a source domain with ample labeled data to a target domain with unlabeled data only. Most existing studies on DA contribute to learning domain-invariant feature representations for both domains by minimizing the domain gap based on convolution-based neural networks. Recently, vision transformers significantly improved performance in multiple vision tasks. Built on vision transformers, in this paper we propose a Bidirectional Cross-Attention Transformer (BCAT) for DA with the aim to improve the performance. In the proposed BCAT, the attention mechanism can extract implicit source and target mixup feature representations to narrow the domain discrepancy. Specifically, in BCAT, we design a weight-sharing quadruple-branch transformer with a bidirectional cross-attention mechanism to learn domain-invariant feature representations. Extensive experiments demonstrate that the proposed BCAT model achieves superior performance on four benchmark datasets over existing state-of-the-art DA methods that are based on convolutions or transformers.

Xiyu Wang, Pengxin Guo, Yu Zhang• 2022

Related benchmarks

TaskDatasetResultRank
Unsupervised Domain AdaptationOffice-Home
Average Accuracy87.4
238
Image ClassificationVisDA 2017 (test)
Class Accuracy (Plane)99.1
83
Unsupervised Domain AdaptationOffice-31
A->W Accuracy99.2
83
Showing 3 of 3 rows

Other info

Follow for update