Learning from Synthetic Data: Addressing Domain Shift for Semantic Segmentation
About
Visual Domain Adaptation is a problem of immense importance in computer vision. Previous approaches showcase the inability of even deep neural networks to learn informative representations across domain shift. This problem is more severe for tasks where acquiring hand labeled data is extremely hard and tedious. In this work, we focus on adapting the representations learned by segmentation networks across synthetic and real domains. Contrary to previous approaches that use a simple adversarial objective or superpixel information to aid the process, we propose an approach based on Generative Adversarial Networks (GANs) that brings the embeddings closer in the learned feature space. To showcase the generality and scalability of our approach, we show that we can achieve state of the art results on two challenging scenarios of synthetic to real domain adaptation. Additional exploratory experiments show that our approach: (1) generalizes to unseen domains and (2) results in improved alignment of source and target distributions.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Semantic segmentation | Cityscapes (val) | mIoU37.1 | 572 | |
| Semantic segmentation | GTA5 → Cityscapes (val) | mIoU37.1 | 533 | |
| Semantic segmentation | SYNTHIA to Cityscapes (val) | Rider IoU15.2 | 435 | |
| Semantic segmentation | Cityscapes GTA5 to Cityscapes adaptation (val) | mIoU (Overall)37.1 | 352 | |
| Semantic segmentation | GTA5 to Cityscapes (test) | mIoU37.1 | 151 | |
| Semantic segmentation | SYNTHIA to Cityscapes | -- | 150 | |
| Semantic segmentation | Cityscapes adaptation from Synthia 1.0 (val) | Person IoU48.2 | 114 | |
| Semantic segmentation | SYNTHIA-to-Cityscapes (SYN2CS) 16 classes (val) | IoU36.1 | 50 | |
| Domain Adaptation Classification | Office-31 (test) | A -> W Accuracy89.5 | 31 |