A Style-Based Generator Architecture for Generative Adversarial Networks
About
We propose an alternative generator architecture for generative adversarial networks, borrowing from style transfer literature. The new architecture leads to an automatically learned, unsupervised separation of high-level attributes (e.g., pose and identity when trained on human faces) and stochastic variation in the generated images (e.g., freckles, hair), and it enables intuitive, scale-specific control of the synthesis. The new generator improves the state-of-the-art in terms of traditional distribution quality metrics, leads to demonstrably better interpolation properties, and also better disentangles the latent factors of variation. To quantify interpolation quality and disentanglement, we propose two new, automated methods that are applicable to any generator architecture. Finally, we introduce a new, highly varied and high-quality dataset of human faces.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Unconditional Image Generation | CIFAR-10 unconditional | -- | 159 | |
| Image Generation | LSUN church | FID3.39 | 95 | |
| Image Generation | LSUN Bedroom 256x256 (test) | FID2.65 | 73 | |
| Class-conditional Image Generation | ImageNet 512x512 | FID2.41 | 72 | |
| Image Generation | LSUN bedroom | FID2.65 | 56 | |
| Image Generation | LSUN Church 256x256 (test) | FID4.21 | 55 | |
| Image Generation | FFHQ | FID3.62 | 52 | |
| Image Generation | CelebA-HQ (test) | FID5.17 | 42 | |
| Unconditional image synthesis | FFHQ 256x256 (test) | FID4.16 | 31 | |
| Text-to-Image Synthesis | MSCOCO | FID13.9 | 31 |