Multiplicative Normalizing Flows for Variational Bayesian Neural Networks
About
We reinterpret multiplicative noise in neural networks as auxiliary random variables that augment the approximate posterior in a variational setting for Bayesian neural networks. We show that through this interpretation it is both efficient and straightforward to improve the approximation by employing normalizing flows while still allowing for local reparametrizations and a tractable lower bound. In experiments we show that with this new approximation we can significantly improve upon classical mean field for Bayesian neural networks on both predictive accuracy as well as predictive uncertainty.
Christos Louizos, Max Welling• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Regression | UCI ENERGY (test) | Negative Log Likelihood3.18 | 42 | |
| Regression | UCI CONCRETE (test) | Neg Log Likelihood-3.35 | 37 | |
| Regression | UCI YACHT (test) | Negative Log Likelihood-1.86 | 33 | |
| Regression | UCI POWER (test) | Negative Log Likelihood-2.86 | 29 | |
| Regression | Energy UCI (test) | RMSE2.38 | 27 | |
| Regression | Boston UCI (test) | RMSE2.98 | 26 | |
| Regression | UCI KIN8NM (test) | -- | 25 | |
| Regression | UCI WINE (test) | Negative Log Likelihood-0.93 | 24 | |
| Regression | UCI NAVAL (test) | Negative Log Likelihood3.96 | 21 | |
| Regression | Concrete UCI (test) | RMSE6.57 | 21 |
Showing 10 of 15 rows