Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Progressive Growing of GANs for Improved Quality, Stability, and Variation

About

We describe a new training methodology for generative adversarial networks. The key idea is to grow both the generator and discriminator progressively: starting from a low resolution, we add new layers that model increasingly fine details as training progresses. This both speeds the training up and greatly stabilizes it, allowing us to produce images of unprecedented quality, e.g., CelebA images at 1024^2. We also propose a simple way to increase the variation in generated images, and achieve a record inception score of 8.80 in unsupervised CIFAR10. Additionally, we describe several implementation details that are important for discouraging unhealthy competition between the generator and discriminator. Finally, we suggest a new metric for evaluating GAN results, both in terms of image quality and variation. As an additional contribution, we construct a higher-quality version of the CelebA dataset.

Tero Karras, Timo Aila, Samuli Laine, Jaakko Lehtinen• 2017

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)--
471
Image GenerationCelebA 64 x 64 (test)
FID16.3
203
Unconditional Image GenerationCIFAR-10 unconditional
FID15.52
159
Image GenerationCIFAR10 32x32 (test)
FID15.52
154
Image GenerationCelebA
FID7.3
110
Image GenerationCIFAR-10 (train/test)
FID18.33
78
Image GenerationLSUN Bedroom 256x256 (test)
FID8.34
73
Image GenerationLSUN bedroom
FID8.3
56
Image GenerationLSUN Church 256x256 (test)
FID6.42
55
Image GenerationCelebA-HQ 256x256
FID8.03
51
Showing 10 of 47 rows

Other info

Code

Follow for update