Masked Autoregressive Flow for Density Estimation
About
Autoregressive models are among the best performing neural density estimators. We describe an approach for increasing the flexibility of an autoregressive model, based on modelling the random numbers that the model uses internally when generating data. By constructing a stack of autoregressive models, each modelling the random numbers of the next model in the stack, we obtain a type of normalizing flow suitable for density estimation, which we call Masked Autoregressive Flow. This type of flow is closely related to Inverse Autoregressive Flow and is a generalization of Real NVP. Masked Autoregressive Flow achieves state-of-the-art performance in a range of general-purpose density estimation tasks.
George Papamakarios, Theo Pavlakou, Iain Murray• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Density Estimation | CIFAR-10 (test) | Bits/dim4.31 | 134 | |
| Density Estimation | MNIST (test) | NLL (bits/dim)1.89 | 56 | |
| Density Estimation | CIFAR-10 | bpd4.31 | 40 | |
| Unconditional Density Estimation | POWER (test) | Average Test Log Likelihood (nats)0.3 | 30 | |
| Density Estimation | GAS d=8; N=1,052,065 (test) | Avg Test Log-Likelihood10.08 | 25 | |
| Density Estimation | BSDS300 (test) | NLL (nats)-156.4 | 25 | |
| Unconditional Density Estimation | HEPMASS (test) | NLL (nats)17.39 | 22 | |
| Unconditional Density Estimation | MINIBOONE (test) | NLL (nats)11.68 | 22 | |
| Audio Generation | LJ Speech (test) | LL Score5.161 | 20 | |
| Density modeling | Nutrient intake data | CvM0.036 | 12 |
Showing 10 of 33 rows