Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Towards Faster and Stabilized GAN Training for High-fidelity Few-shot Image Synthesis

About

Training Generative Adversarial Networks (GAN) on high-fidelity images usually requires large-scale GPU-clusters and a vast number of training images. In this paper, we study the few-shot image synthesis task for GAN with minimum computing cost. We propose a light-weight GAN structure that gains superior quality on 1024*1024 resolution. Notably, the model converges from scratch with just a few hours of training on a single RTX-2080 GPU, and has a consistent performance, even with less than 100 training samples. Two technique designs constitute our work, a skip-layer channel-wise excitation module and a self-supervised discriminator trained as a feature-encoder. With thirteen datasets covering a wide variety of image domains (The datasets and code are available at: https://github.com/odegeasslbc/FastGAN-pytorch), we show our model's superior performance compared to the state-of-the-art StyleGAN2, when data and computing budget are limited.

Bingchen Liu, Yizhe Zhu, Kunpeng Song, Ahmed Elgammal• 2021

Related benchmarks

TaskDatasetResultRank
Image GenerationLSUN church
FID8.43
95
Image GenerationLSUN bedroom
FID3
56
Image GenerationFFHQ
FID12.38
52
Image GenerationCIFAR-10 unconditional (test)
FID34.5
39
Image GenerationPanda 100-shot (train)
FID10.03
28
Image GenerationGrumpy cat 100-shot (train)
FID26.65
28
Image GenerationObama 100-shot (train)
FID41.05
28
Image GenerationFFHQ
FID12.69
22
Image GenerationAnimalFace Dog
FID62.11
21
Image GenerationAnimalFace Dog standard (train)
FID50.66
20
Showing 10 of 69 rows

Other info

Code

Follow for update