T-3DGS: Removing Transient Objects for 3D Scene Reconstruction
About
Transient objects in video sequences can significantly degrade the quality of 3D scene reconstructions. To address this challenge, we propose T-3DGS, a novel framework that robustly filters out transient distractors during 3D reconstruction using Gaussian Splatting. Our framework consists of two steps. First, we employ an unsupervised classification network that distinguishes transient objects from static scene elements by leveraging their distinct training dynamics within the reconstruction process. Second, we refine these initial detections by integrating an off-the-shelf segmentation method with a bidirectional tracking module, which together enhance boundary accuracy and temporal coherence. Evaluations on both sparsely and densely captured video datasets demonstrate that T-3DGS significantly outperforms state-of-the-art approaches, enabling high-fidelity 3D reconstructions in challenging, real-world scenarios.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Novel View Synthesis | D-RE10K static regions only (test) | PSNR18.7 | 26 | |
| Novel View Synthesis | D-RE10K-iPhone full-image fidelity (test) | PSNR16.95 | 26 | |
| Novel View Synthesis | RobustNeRF Baby Yoda scene | LPIPS0.095 | 20 | |
| Novel View Synthesis | RobustNeRF Statue | PSNR22.68 | 17 | |
| Novel View Synthesis | RobustNeRF Android | PSNR24.67 | 17 | |
| Novel View Synthesis | RobustNeRF Crab | PSNR33.11 | 16 | |
| Novel View Synthesis | RobustNeRF Avg. | PSNR28.37 | 12 | |
| Novel View Synthesis | On-the-go Dataset | PSNR (Mountain)20.5 | 12 | |
| 3D Scene Reconstruction | RobustNeRF and On-the-go Average (test) | Average Training Time (min)26.75 | 6 |