Deep AutoRegressive Networks
About
We introduce a deep, generative autoencoder capable of learning hierarchies of distributed representations from data. Successive deep stochastic hidden layers are equipped with autoregressive connections, which enable the model to be sampled from quickly and exactly via ancestral sampling. We derive an efficient approximate parameter estimation method based on the minimum description length (MDL) principle, which can be seen as maximising a variational lower bound on the log-likelihood, with a feedforward neural network implementing approximate inference. We demonstrate state-of-the-art generative performance on a number of classic data sets: several UCI data sets, MNIST and Atari 2600 games.
Karol Gregor, Ivo Danihelka, Andriy Mnih, Charles Blundell, Daan Wierstra• 2013
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Density Estimation | binarized MNIST 28x28 (test) | Test LogL-84.13 | 44 | |
| Density Estimation | Ocr-letters (test) | -- | 19 | |
| Generative Modeling | MNIST Binary (test) | NLL (nats)84.13 | 13 | |
| Density Estimation | Adult UCI repository (test) | -- | 9 | |
| Density Estimation | Connect4 (test) | -- | 9 | |
| Density Estimation | dna (test) | -- | 9 | |
| Density Estimation | Mushrooms (test) | -- | 9 | |
| Density Estimation | Web (test) | -- | 9 | |
| Distribution Estimation | RCV1 (test) | Negative Log-Likelihood46.1 | 8 | |
| Distribution Estimation | NIPS-0-12 (test) | Negative Log-Likelihood274.7 | 8 |
Showing 10 of 10 rows