Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Self-Attention Generative Adversarial Networks

About

In this paper, we propose the Self-Attention Generative Adversarial Network (SAGAN) which allows attention-driven, long-range dependency modeling for image generation tasks. Traditional convolutional GANs generate high-resolution details as a function of only spatially local points in lower-resolution feature maps. In SAGAN, details can be generated using cues from all feature locations. Moreover, the discriminator can check that highly detailed features in distant portions of the image are consistent with each other. Furthermore, recent work has shown that generator conditioning affects GAN performance. Leveraging this insight, we apply spectral normalization to the GAN generator and find that this improves training dynamics. The proposed SAGAN achieves the state-of-the-art results, boosting the best published Inception score from 36.8 to 52.52 and reducing Frechet Inception distance from 27.62 to 18.65 on the challenging ImageNet dataset. Visualization of the attention layers shows that the generator leverages neighborhoods that correspond to object shapes rather than local regions of fixed shape.

Han Zhang, Ian Goodfellow, Dimitris Metaxas, Augustus Odena• 2018

Related benchmarks

TaskDatasetResultRank
Image GenerationCIFAR-10 (test)
FID13.4
471
Image GenerationCIFAR-10--
178
Class-conditional Image GenerationImageNet
FID18.65
132
Image GenerationLSUN church
FID6.15
95
Image GenerationLSUN bedroom
FID14.06
56
Image GenerationFFHQ
FID16.21
52
Image GenerationImageNet 128x128
FID18.28
51
Across-modality synthesis (T2-weighted MRI to CT)Pelvic MRI-CT dataset (test)
PSNR27.61
42
Image GenerationTiny-ImageNet--
34
Image GenerationCIFAR10 (train)
FID0.45
32
Showing 10 of 64 rows

Other info

Code

Follow for update