Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Discrete Variational Autoencoders

About

Probabilistic models with discrete latent variables naturally capture datasets composed of discrete classes. However, they are difficult to train efficiently, since backpropagation through discrete variables is generally not possible. We present a novel method to train a class of probabilistic models with discrete latent variables using the variational autoencoder framework, including backpropagation through the discrete latent variables. The associated class of probabilistic models comprises an undirected discrete component and a directed hierarchical continuous component. The discrete component captures the distribution over the disconnected smooth manifolds induced by the continuous component. As a result, this class of models efficiently learns both the class of objects in an image, and their specific realization in pixels, from unsupervised data, and outperforms state-of-the-art methods on the permutation-invariant MNIST, Omniglot, and Caltech-101 Silhouettes datasets.

Jason Tyler Rolfe• 2016

Related benchmarks

TaskDatasetResultRank
Log-likelihood estimationMNIST dynamically binarized (test)
Log-Likelihood80.04
48
Density Estimationbinarized MNIST 28x28 (test)
Test LogL81.01
44
Image ModelingOmniglot (test)
NLL97.43
27
Density EstimationOMNIGLOT dynamically binarized (test)
NLL97.43
16
Likelihood EstimationMNIST
NLL80.15
7
Showing 5 of 5 rows

Other info

Follow for update