GO-SLAM: Global Optimization for Consistent 3D Instant Reconstruction
About
Neural implicit representations have recently demonstrated compelling results on dense Simultaneous Localization And Mapping (SLAM) but suffer from the accumulation of errors in camera tracking and distortion in the reconstruction. Purposely, we present GO-SLAM, a deep-learning-based dense visual SLAM framework globally optimizing poses and 3D reconstruction in real-time. Robust pose estimation is at its core, supported by efficient loop closing and online full bundle adjustment, which optimize per frame by utilizing the learned global geometry of the complete history of input frames. Simultaneously, we update the implicit and continuous surface representation on-the-fly to ensure global consistency of 3D reconstruction. Results on various synthetic and real-world datasets demonstrate that GO-SLAM outperforms state-of-the-art approaches at tracking robustness and reconstruction accuracy. Furthermore, GO-SLAM is versatile and can run with monocular, stereo, and RGB-D input.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Camera pose estimation | ScanNet | -- | 119 | |
| Visual-Inertial Odometry | EuRoC (All sequences) | MH1 Error0.016 | 62 | |
| Camera Tracking | Replica | Rotation Error (rm-0)0.34 | 38 | |
| Visual Odometry | TUM-RGBD | freiburg1/desk2 Error0.028 | 37 | |
| Absolute Trajectory Estimation | TUM RGB-D | Desk Error0.015 | 36 | |
| Visual SLAM | TUM RGB-D fr1 desk | ATE RMSE (cm)2.119 | 24 | |
| Visual SLAM | TUM RGB-D fr2 xyz | Translation RMSE (m)0.2858 | 21 | |
| Tracking | ScanNet | ATE RMSE (Seq 00)5.4 | 18 | |
| Novel View Synthesis | SeaThru-NeRF (J.G.-RedSea) | PSNR15.72 | 18 | |
| Novel View Synthesis | SeaThru-NeRF Panama | PSNR16.63 | 18 |