Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

High-Fidelity Image Generation With Fewer Labels

About

Deep generative models are becoming a cornerstone of modern machine learning. Recent work on conditional generative adversarial networks has shown that learning complex, high-dimensional distributions over natural images is within reach. While the latest models are able to generate high-fidelity, diverse natural images at high resolution, they rely on a vast quantity of labeled data. In this work we demonstrate how one can benefit from recent work on self- and semi-supervised learning to outperform the state of the art on both unsupervised ImageNet synthesis, as well as in the conditional setting. In particular, the proposed approach is able to match the sample quality (as measured by FID) of the current state-of-the-art conditional model BigGAN on ImageNet using only 10% of the labels and outperform it using 20% of the labels.

Mario Lucic, Michael Tschannen, Marvin Ritter, Xiaohua Zhai, Olivier Bachem, Sylvain Gelly• 2019

Related benchmarks

TaskDatasetResultRank
Image GenerationImageNet (val)
FID22
198
Class-conditional image synthesisTiny ImageNet (train test)
FID20.95
60
Image GenerationImageNet 128x128
FID25.3
51
Image GenerationImageNet (train val)
Precision69.6
17
Scene GenerationCOCO Stuff (val)
FID46.9
14
Scene GenerationCOCO-Stuff unseen (eval)
FID60.9
14
Scene GenerationCOCO-Stuff seen (val)
FID103.8
14
Class-conditional image synthesisImageNet
FID180.3
13
Scene GenerationCOCO-Stuff (train)
FID17.9
12
Image GenerationImageNet 128x128 (train val)--
8
Showing 10 of 10 rows

Other info

Code

Follow for update