Learning Unsupervised Cross-domain Image-to-Image Translation Using a Shared Discriminator
About
Unsupervised image-to-image translation is used to transform images from a source domain to generate images in a target domain without using source-target image pairs. Promising results have been obtained for this problem in an adversarial setting using two independent GANs and attention mechanisms. We propose a new method that uses a single shared discriminator between the two GANs, which improves the overall efficacy. We assess the qualitative and quantitative results on image transfiguration, a cross-domain translation task, in a setting where the target domain shares similar semantics to the source domain. Our results indicate that even without adding attention mechanisms, our method performs at par with attention-based methods and generates images of comparable quality.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unpaired Image-to-Image Translation (Apple to Orange) | Apple2Orange (test) | KID0.044 | 8 | |
| Unpaired Image-to-Image Translation (Zebra to Horse) | Horse2Zebra (test) | KID (x100)5.8 | 8 | |
| Unpaired Image-to-Image Translation (Horse to Zebra) | Horse2Zebra (test) | KID (x100)3 | 8 | |
| Unpaired Image-to-Image Translation (Orange to Apple) | Apple2Orange (test) | KID0.051 | 8 | |
| Image-to-Image Translation | Horse2Zebra (test) | FID92.91 | 6 |