Emerging Convolutions for Generative Normalizing Flows
About
Generative flows are attractive because they admit exact likelihood optimization and efficient image synthesis. Recently, Kingma & Dhariwal (2018) demonstrated with Glow that generative flows are capable of generating high quality images. We generalize the 1 x 1 convolutions proposed in Glow to invertible d x d convolutions, which are more flexible since they operate on both channel and spatial axes. We propose two methods to produce invertible convolutions that have receptive fields identical to standard convolutions: Emerging convolutions are obtained by chaining specific autoregressive convolutions, and periodic convolutions are decoupled in the frequency domain. Our experiments show that the flexibility of d x d convolutions significantly improves the performance of generative flow models on galaxy images, CIFAR10 and ImageNet.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Density Estimation | CIFAR-10 (test) | Bits/dim3.34 | 134 | |
| Density Estimation | ImageNet 32x32 (test) | Bits per Sub-pixel4.09 | 66 | |
| Density Estimation | ImageNet 64x64 (test) | Bits Per Sub-Pixel3.81 | 62 | |
| Unconditional Image Generation | CIFAR10 | BPD3.34 | 33 | |
| Unconditional Image Generation | ImageNet-32 | BPD4.09 | 31 | |
| Unconditional Image Generation | ImageNet 64 | BPD3.81 | 22 | |
| Generative Modeling | ImageNet 64x64 downsampled | Bits Per Dimension3.81 | 13 | |
| Image Generation | MNIST (test) | -- | 13 | |
| Generative Modeling | MNIST | -- | 10 | |
| Generative Modeling | ImageNet 32x32 | BPD4.09 | 4 |