Image Generators with Conditionally-Independent Pixel Synthesis
About
Existing image generator networks rely heavily on spatial convolutions and, optionally, self-attention blocks in order to gradually synthesize images in a coarse-to-fine manner. Here, we present a new architecture for image generators, where the color value at each pixel is computed independently given the value of a random latent vector and the coordinate of that pixel. No spatial convolutions or similar operations that propagate information across pixels are involved during the synthesis. We analyze the modeling capabilities of such generators when trained in an adversarial fashion, and observe the new generators to achieve similar generation quality to state-of-the-art convolutional generators. We also investigate several interesting properties unique to the new architecture.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Generation | LSUN church | FID2.92 | 95 | |
| Unconditional Image Generation | FFHQ 256x256 | FID4.38 | 64 | |
| Image Generation | LSUN Church 256x256 (test) | FID2.92 | 55 | |
| Image Generation | FFHQ | FID4.38 | 52 | |
| Image Generation | FFHQ 256x256 (test) | FID4.38 | 30 | |
| Multi-scale Image Generation | FFHQ 1024x1024 (test) | Self-SSIM0.9991 | 24 | |
| Image Generation | FFHQ (test) | FID4.38 | 21 | |
| Unconditional Image Generation | LSUN Church 256x256 | FID2.92 | 14 | |
| Unconditional image synthesis | FFHQ 1024 | FID10.07 | 12 | |
| Unconditional image synthesis | FFHQ 512 | FID6.18 | 3 |