SGS-SLAM: Semantic Gaussian Splatting For Neural Dense SLAM
About
We present SGS-SLAM, the first semantic visual SLAM system based on Gaussian Splatting. It incorporates appearance, geometry, and semantic features through multi-channel optimization, addressing the oversmoothing limitations of neural implicit SLAM systems in high-quality rendering, scene understanding, and object-level geometry. We introduce a unique semantic feature loss that effectively compensates for the shortcomings of traditional depth and color losses in object optimization. Through a semantic-guided keyframe selection strategy, we prevent erroneous reconstructions caused by cumulative errors. Extensive experiments demonstrate that SGS-SLAM delivers state-of-the-art performance in camera pose estimation, map reconstruction, precise semantic segmentation, and object-level geometric accuracy, while ensuring real-time rendering capabilities.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Novel View Synthesis | ScanNet | PSNR22.27 | 58 | |
| Photometric Reconstruction | Replica (room1) | PSNR31.6 | 8 | |
| Photometric Reconstruction | Replica | PSNR32.11 | 8 | |
| Photometric Reconstruction | Replica (room0) | PSNR29.83 | 8 | |
| Photometric Reconstruction | Replica (room2) | PSNR32.68 | 8 | |
| Photometric Reconstruction | Replica (office0) | PSNR36.75 | 8 | |
| Photometric Reconstruction | Replica (office1) | PSNR37.28 | 8 | |
| Photometric Reconstruction | Replica (office4) | PSNR30.36 | 8 | |
| Photometric Reconstruction | Replica office2 | PSNR29.72 | 8 | |
| Photometric Reconstruction | Replica office3 | PSNR28.63 | 8 |