Kalman Bayesian Neural Networks for Closed-form Online Learning
About
Compared to point estimates calculated by standard neural networks, Bayesian neural networks (BNN) provide probability distributions over the output predictions and model parameters, i.e., the weights. Training the weight distribution of a BNN, however, is more involved due to the intractability of the underlying Bayesian inference problem and thus, requires efficient approximations. In this paper, we propose a novel approach for BNN learning via closed-form Bayesian inference. For this purpose, the calculation of the predictive distribution of the output and the update of the weight distribution are treated as Bayesian filtering and smoothing problems, where the weights are modeled as Gaussian random variables. This allows closed-form expressions for training the network's parameters in a sequential/online fashion without gradient descent. We demonstrate our method on several UCI datasets and compare it to the state of the art.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Regression | Yacht | RMSE14.51 | 49 | |
| Regression | UCI ENERGY (test) | Negative Log Likelihood8.24 | 42 | |
| Regression | UCI CONCRETE (test) | Neg Log Likelihood8.69 | 37 | |
| Regression | UCI YACHT (test) | Negative Log Likelihood4.37 | 33 | |
| Regression | UCI KIN8NM (test) | NLL-0.41 | 25 | |
| Regression | UCI WINE (test) | Negative Log Likelihood4.97 | 24 | |
| Regression | UCI NAVAL (test) | Negative Log Likelihood9.68 | 21 | |
| Regression | Energy | Avg NLL Relative Percentage0.00e+0 | 8 | |
| Regression | Wine | Avg NLL Relative Percentage6 | 8 | |
| Regression | Concrete | Avg NLL Relative %6 | 8 |