Activation-Space Uncertainty Quantification for Pretrained Networks
About
Reliable uncertainty estimates are crucial for deploying pretrained models; yet, many strong methods for quantifying uncertainty require retraining, Monte Carlo sampling, or expensive second-order computations and may alter a frozen backbone's predictions. To address this, we introduce Gaussian Process Activations (GAPA), a post-hoc method that shifts Bayesian modeling from weights to activations. GAPA replaces standard nonlinearities with Gaussian-process activations whose posterior mean exactly matches the original activation, preserving the backbone's point predictions by construction while providing closed-form epistemic variances in activation space. To scale to modern architectures, we use a sparse variational inducing-point approximation over cached training activations, combined with local k-nearest-neighbor subset conditioning, enabling deterministic single-pass uncertainty propagation without sampling, backpropagation, or second-order information. Across regression, classification, image segmentation, and language modeling, GAPA matches or outperforms strong post-hoc baselines in calibration and out-of-distribution detection while remaining efficient at test time.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | FashionMNIST (test) | Accuracy85.9 | 218 | |
| Out-of-Distribution Detection | CIFAR-10 | AUROC95.3 | 36 | |
| Image Classification | CIFAR-10 | Accuracy0.944 | 36 | |
| OOD Detection | CIFAR-10 | OOD AUROC0.953 | 24 | |
| Image Classification | CIFAR-10 | NLL0.23 | 24 | |
| Regression | Airline | NLL4.946 | 22 | |
| Classification and Uncertainty Quantification | MNIST (test) | Accuracy97.8 | 15 | |
| Uncertainty Quantification | CIFAR-10 (test) | Accuracy93.5 | 14 | |
| Out-of-Distribution Detection | FMNIST | OOD Score97.3 | 13 | |
| Out-of-Distribution Detection | MNIST | OOD Score96 | 13 |