Label-Driven Reconstruction for Domain Adaptation in Semantic Segmentation
About
Unsupervised domain adaptation enables to alleviate the need for pixel-wise annotation in the semantic segmentation. One of the most common strategies is to translate images from the source domain to the target domain and then align their marginal distributions in the feature space using adversarial learning. However, source-to-target translation enlarges the bias in translated images and introduces extra computations, owing to the dominant data size of the source domain. Furthermore, consistency of the joint distribution in source and target domains cannot be guaranteed through global feature alignment. Here, we present an innovative framework, designed to mitigate the image translation bias and align cross-domain features with the same category. This is achieved by 1) performing the target-to-source translation and 2) reconstructing both source and target images from their predicted labels. Extensive experiments on adapting from synthetic to real urban scene understanding demonstrate that our framework competes favorably against existing state-of-the-art methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic segmentation | GTA5 → Cityscapes (val) | mIoU49.5 | 533 | |
| Semantic segmentation | SYNTHIA to Cityscapes (val) | Rider IoU73.2 | 435 | |
| Semantic segmentation | Cityscapes GTA5 to Cityscapes adaptation (val) | mIoU (Overall)49.5 | 352 | |
| Semantic segmentation | SYNTHIA to Cityscapes | Road IoU85.1 | 150 | |
| Semantic segmentation | Synthia to Cityscapes (test) | Road IoU85.1 | 138 | |
| Semantic segmentation | Cityscapes (val) | mIoU49.5 | 133 | |
| Semantic segmentation | GTA5 to Cityscapes 1.0 (val) | Road IoU90.8 | 98 | |
| Semantic segmentation | Cityscapes (val) | IoU (road)90.8 | 37 |