Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Ladder Variational Autoencoders

About

Variational Autoencoders are powerful models for unsupervised learning. However deep models with several layers of dependent stochastic variables are difficult to train which limits the improvements obtained using these highly expressive models. We propose a new inference model, the Ladder Variational Autoencoder, that recursively corrects the generative distribution by a data dependent approximate likelihood in a process resembling the recently proposed Ladder Network. We show that this model provides state of the art predictive log-likelihood and tighter log-likelihood lower bound compared to the purely bottom-up inference in layered Variational Autoencoders and other generative models. We provide a detailed analysis of the learned hierarchical latent representation and show that our new inference model is qualitatively different and utilizes a deeper more distributed hierarchy of latent variables. Finally, we observe that batch normalization and deterministic warm-up (gradually turning on the KL-term) are crucial for training variational models with many stochastic layers.

Casper Kaae S{\o}nderby, Tapani Raiko, Lars Maal{\o}e, S{\o}ren Kaae S{\o}nderby, Ole Winther• 2016

Related benchmarks

TaskDatasetResultRank
Log-likelihood estimationMNIST dynamically binarized (test)
Log-Likelihood81.74
48
Generative ModelingMNIST (test)--
35
Image ModelingOmniglot (test)
NLL102.1
27
Density EstimationOMNIGLOT dynamically binarized (test)
NLL102.1
16
Generative ModelingMNIST permutation-invariant (test)
Log Likelihood-81.74
10
Generative ModelingOmniglot (test)--
8
Showing 6 of 6 rows

Other info

Follow for update