TriaGS: Differentiable Triangulation-Guided Geometric Consistency for 3D Gaussian Splatting
About
3D Gaussian Splatting is crucial for real-time novel view synthesis due to its efficiency and ability to render photorealistic images. However, building a 3D Gaussian is guided solely by photometric loss, which can result in inconsistencies in reconstruction. This under-constrained process often results in "floater" artifacts and unstructured geometry, preventing the extraction of high-fidelity surfaces. To address this issue, our paper introduces a novel method that improves reconstruction by enforcing global geometry consistency through constrained multi-view triangulation. Our approach aims to achieve a consensus on 3D representation in the physical world by utilizing various estimated views. We optimize this process by penalizing the deviation of a rendered 3D point from a robust consensus point, which is re-triangulated from a bundle of neighboring views in a self-supervised fashion. We demonstrate the effectiveness of our method across multiple datasets, achieving state-of-the-art results. On the DTU dataset, our method attains a mean Chamfer Distance of 0.50 mm, outperforming comparable explicit methods. We will make our code open-source to facilitate community validation and ensure reproducibility.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Novel View Synthesis | MipNeRF 360 Outdoor | PSNR24.95 | 112 | |
| Novel View Synthesis | MipNeRF 360 Indoor | PSNR30.89 | 108 | |
| Surface Reconstruction | DTU | Scan 24 Metric Value0.35 | 34 | |
| Surface Reconstruction | NeRF Synthetic | Chair Value0.5 | 11 | |
| Surface Reconstruction | Tanks&Temples (test) | Barn F1 Score62 | 6 |