Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Few-Shot Unsupervised Image-to-Image Translation

About

Unsupervised image-to-image translation methods learn to map images in a given class to an analogous image in a different class, drawing on unstructured (non-registered) datasets of images. While remarkably successful, current methods require access to many images in both source and destination classes at training time. We argue this greatly limits their use. Drawing inspiration from the human capability of picking up the essence of a novel object from a small number of examples and generalizing from there, we seek a few-shot, unsupervised image-to-image translation algorithm that works on previously unseen target classes that are specified, at test time, only by a few example images. Our model achieves this few-shot generation capability by coupling an adversarial training scheme with a novel network design. Through extensive experimental validation and comparisons to several baseline methods on benchmark datasets, we verify the effectiveness of the proposed framework. Our implementation and datasets are available at https://github.com/NVlabs/FUNIT .

Ming-Yu Liu, Xun Huang, Arun Mallya, Tero Karras, Timo Aila, Jaakko Lehtinen, Jan Kautz• 2019

Related benchmarks

TaskDatasetResultRank
Font GenerationCollected Chinese font dataset Unseen styles and Unseen contents
SSIM0.7074
13
Font GenerationCollected Chinese font dataset (Seen styles and Unseen contents)
SSIM0.7269
8
Stylized Face GenerationAAHQ
FID87.71
7
Font GenerationChinese Font Dataset (Seen Fonts)
L1 Error0.0859
7
Font GenerationUFSC Unseen Fonts Seen Characters
L1 Loss0.148
6
Font GenerationUFUC Unseen Fonts Unseen Characters
L1 Loss0.152
6
Latent-guided stylized face generationAAHQ (test)
FID177
6
Showing 7 of 7 rows

Other info

Follow for update