Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Large Scale GAN Training for High Fidelity Natural Image Synthesis

About

Despite recent progress in generative image modeling, successfully generating high-resolution, diverse samples from complex datasets such as ImageNet remains an elusive goal. To this end, we train Generative Adversarial Networks at the largest scale yet attempted, and study the instabilities specific to such scale. We find that applying orthogonal regularization to the generator renders it amenable to a simple "truncation trick," allowing fine control over the trade-off between sample fidelity and variety by reducing the variance of the Generator's input. Our modifications lead to models which set the new state of the art in class-conditional image synthesis. When trained on ImageNet at 128x128 resolution, our models (BigGANs) achieve an Inception Score (IS) of 166.5 and Frechet Inception Distance (FID) of 7.4, improving over the previous best IS of 52.52 and FID of 18.6.

Andrew Brock, Jeff Donahue, Karen Simonyan• 2018

Related benchmarks

TaskDatasetResultRank
Image ClassificationImageNet-1K 1.0 (val)
Top-1 Accuracy64
1952
Class-conditional Image GenerationImageNet 256x256
Inception Score (IS)317
815
Image GenerationCIFAR-10 (test)
FID6.04
483
Class-conditional Image GenerationImageNet 256x256 (val)
FID6.95
427
Image GenerationImageNet 256x256
IS224.5
359
Class-conditional Image GenerationImageNet 256x256 (train)
IS224.5
345
Image GenerationImageNet 256x256 (val)
FID6.95
340
Unconditional Image GenerationCIFAR-10
FID8.51
240
Image GenerationImageNet 512x512 (val)
FID-50K8.43
219
Class-conditional Image GenerationImageNet 256x256 (test)
FID6.95
208
Showing 10 of 151 rows
...

Other info

Code

Follow for update