Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Improved Training of Wasserstein GANs

About

Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but sometimes can still generate only low-quality samples or fail to converge. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to undesired behavior. We propose an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input. Our proposed method performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning, including 101-layer ResNets and language models over discrete data. We also achieve high quality generations on CIFAR-10 and LSUN bedrooms.

Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron Courville• 2017

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID14.1
483
Image GenerationCelebA 64 x 64 (test)
FID12.89
208
Image GenerationCIFAR10 32x32 (test)
FID40.2
183
Image GenerationCIFAR-10
Inception Score6.51
178
Unconditional Image GenerationCIFAR-10 unconditional
FID39.68
165
Image DehazingSOTS (test)
PSNR25.28
161
Image DerainingRain100L (test)
PSNR32.87
161
DehazingSOTS
PSNR20.92
154
Image GenerationCelebA
FID15.23
110
Conditional Image GenerationCIFAR10 (test)
Fréchet Inception Distance24.8
92
Showing 10 of 72 rows
...

Other info

Follow for update