Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LOGAN: Latent Optimisation for Generative Adversarial Networks

About

Training generative adversarial networks requires balancing of delicate adversarial dynamics. Even with careful tuning, training may diverge or end up in a bad equilibrium with dropped modes. In this work, we improve CS-GAN with natural gradient-based latent optimisation and show that it improves adversarial dynamics by enhancing interactions between the discriminator and the generator. Our experiments demonstrate that latent optimisation can significantly improve GAN training, obtaining state-of-the-art performance for the ImageNet ($128 \times 128$) dataset. Our model achieves an Inception Score (IS) of $148$ and an Fr\'echet Inception Distance (FID) of $3.4$, an improvement of $17\%$ and $32\%$ in IS and FID respectively, compared with the baseline BigGAN-deep model with the same architecture and number of parameters.

Yan Wu, Jeff Donahue, David Balduzzi, Karen Simonyan, Timothy Lillicrap• 2019

Related benchmarks

TaskDatasetResultRank
Class-conditional Image GenerationImageNet 128x128--
27
Image SynthesisImageNet 128x128
FID3.36
5
Showing 2 of 2 rows

Other info

Follow for update