Online Structured Laplace Approximations For Overcoming Catastrophic Forgetting
About
We introduce the Kronecker factored online Laplace approximation for overcoming catastrophic forgetting in neural networks. The method is grounded in a Bayesian online learning framework, where we recursively approximate the posterior after every task with a Gaussian, leading to a quadratic penalty on changes to the weights. The Laplace approximation requires calculating the Hessian around a mode, which is typically intractable for modern architectures. In order to make our method scalable, we leverage recent block-diagonal Kronecker factored approximations to the curvature. Our algorithm achieves over 90% test accuracy across a sequence of 50 instantiations of the permuted MNIST dataset, substantially outperforming related methods for overcoming catastrophic forgetting.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | CIFAR-100 Split | Accuracy76.61 | 61 | |
| Image Classification | MNIST Split | Test Accuracy99.04 | 24 | |
| Fake Image Detection | Deepfake | AA98.37 | 9 | |
| Fake Image Detection | VQDM | AA85.92 | 9 | |
| Fake Image Detection | StyleGAN2 | AA Score94.98 | 9 | |
| Fake Image Detection | SD v2.1 | AA69.61 | 9 | |
| Fake Image Detection | SDXL v1.0 | AA73.33 | 9 | |
| Fake Image Detection | SD v3.0 | AA79.08 | 9 | |
| Fake Image Detection | ProGAN | Accuracy99.99 | 9 | |
| Fake Image Detection | BigGAN | AA90.25 | 9 |