Predictive Photometric Uncertainty in Gaussian Splatting for Novel View Synthesis
About
Recent advances in 3D Gaussian Splatting have enabled impressive photorealistic novel view synthesis. However, to transition from a pure rendering engine to a reliable spatial map for autonomous agents and safety-critical applications, knowing where the representation is uncertain is as important as the rendering fidelity itself. We bridge this critical gap by introducing a lightweight, plug-and-play framework for pixel-wise, view-dependent predictive uncertainty estimation. Our post-hoc method formulates uncertainty as a Bayesian-regularized linear least-squares optimization over reconstruction residuals. This architecture-agnostic approach extracts a per-primitive uncertainty channel without modifying the underlying scene representation or degrading baseline visual fidelity. Crucially, we demonstrate that providing this actionable reliability signal successfully translates 3D Gaussian splatting into a trustworthy spatial map, further improving state-of-the-art performance across three critical downstream perception tasks: active view selection, pose-agnostic scene change detection, and pose-agnostic anomaly detection.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Scene Change Detection | PASLCD | mIoU49.8 | 8 | |
| Active View Selection | Mip-NeRF360 | PSNR20.676 | 4 | |
| Anomaly Detection | MAD-Real (test) | AUROC95.6 | 4 | |
| Uncertainty Estimation for Novel View Synthesis | Mip-NeRF 360 | AUSE (L1)0.328 | 4 | |
| Uncertainty Estimation for Novel View Synthesis | Tanks&Temples | AUSE (L1)0.299 | 4 | |
| Uncertainty Estimation for Novel View Synthesis | Deep Blending | AUSE (L1)0.376 | 4 |