Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Locally Masked Convolution for Autoregressive Models

About

High-dimensional generative models have many applications including image compression, multimedia generation, anomaly detection and data completion. State-of-the-art estimators for natural images are autoregressive, decomposing the joint distribution over pixels into a product of conditionals parameterized by a deep neural network, e.g. a convolutional neural network such as the PixelCNN. However, PixelCNNs only model a single decomposition of the joint, and only a single generation order is efficient. For tasks such as image completion, these models are unable to use much of the observed context. To generate data in arbitrary orders, we introduce LMConv: a simple modification to the standard 2D convolution that allows arbitrary masks to be applied to the weights at each location in the image. Using LMConv, we learn an ensemble of distribution estimators that share parameters but differ in generation order, achieving improved performance on whole-image density estimation (2.89 bpd on unconditional CIFAR10), as well as globally coherent image completions. Our code is available at https://ajayjain.github.io/lmconv.

Ajay Jain, Pieter Abbeel, Deepak Pathak• 2020

Related benchmarks

TaskDatasetResultRank
Density EstimationCIFAR-10 (test)--
134
Density Estimationbinarized MNIST 28x28 (test)--
44
Density EstimationCIFAR-10
bpd2.89
40
Generative ModelingCIFAR-10 8-bit color (test)
Bits per Dimension2.89
15
Generative ModelingDynamically binarized MNIST (test)--
13
Density EstimationCelebAHQ 256 x 256 5-bits
NLL (bits/dim)0.74
8
Density EstimationMNIST
NLL (nats)77.58
5
Density EstimationGrayscale MNIST 28x28 (test)
NLL (bpd)0.65
4
Showing 8 of 8 rows

Other info

Code

Follow for update