Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

U-GAT-IT: Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization for Image-to-Image Translation

About

We propose a novel method for unsupervised image-to-image translation, which incorporates a new attention module and a new learnable normalization function in an end-to-end manner. The attention module guides our model to focus on more important regions distinguishing between source and target domains based on the attention map obtained by the auxiliary classifier. Unlike previous attention-based method which cannot handle the geometric changes between domains, our model can translate both images requiring holistic changes and images requiring large shape changes. Moreover, our new AdaLIN (Adaptive Layer-Instance Normalization) function helps our attention-guided model to flexibly control the amount of change in shape and texture by learned parameters depending on datasets. Experimental results show the superiority of the proposed method compared to the existing state-of-the-art models with a fixed network architecture and hyper-parameters. Our code and datasets are available at https://github.com/taki0112/UGATIT or https://github.com/znxlwm/UGATIT-pytorch.

Junho Kim, Minjae Kim, Hyeonwoo Kang, Kwanghee Lee• 2019

Related benchmarks

TaskDatasetResultRank
Image-to-Image TranslationRetinal Fundus-to-Angiogram (test)
FID24.5
42
Image-to-Image Translationselfie2anime
KID0.1161
11
Image-to-Image Translationanime2selfie
KID0.1152
10
Image-to-Image Translationportrait2photo
KID1.69
10
Sketch-to-Photo GenerationChair V2
FID107.2
8
Sketch-to-Photo GenerationHandbag
FID127.5
8
Sketch-to-Photo GenerationShoe V2
FID76.89
8
Image-to-Image Translationselfie2anime
Preference Score73.15
7
Image-to-Image Translationhorse2zebra
Preference Score73.56
7
Image-to-Image TranslationFundus-to-Angiography Original (test)
FID24.5
7
Showing 10 of 43 rows

Other info

Code

Follow for update