GO-SLAM: Global Optimization for Consistent 3D Instant Reconstruction
About
Neural implicit representations have recently demonstrated compelling results on dense Simultaneous Localization And Mapping (SLAM) but suffer from the accumulation of errors in camera tracking and distortion in the reconstruction. Purposely, we present GO-SLAM, a deep-learning-based dense visual SLAM framework globally optimizing poses and 3D reconstruction in real-time. Robust pose estimation is at its core, supported by efficient loop closing and online full bundle adjustment, which optimize per frame by utilizing the learned global geometry of the complete history of input frames. Simultaneously, we update the implicit and continuous surface representation on-the-fly to ensure global consistency of 3D reconstruction. Results on various synthetic and real-world datasets demonstrate that GO-SLAM outperforms state-of-the-art approaches at tracking robustness and reconstruction accuracy. Furthermore, GO-SLAM is versatile and can run with monocular, stereo, and RGB-D input.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Camera pose estimation | ScanNet | -- | 61 | |
| Visual-Inertial Odometry | EuRoC (All sequences) | MH1 Error0.016 | 51 | |
| Visual Odometry | TUM-RGBD | freiburg1/xyz Error0.01 | 34 | |
| Absolute Trajectory Estimation | TUM RGB-D | Desk Error0.015 | 23 | |
| Visual SLAM | TUM RGB-D fr1 desk | ATE RMSE (cm)2.119 | 21 | |
| Visual SLAM | TUM RGB-D fr2 xyz | Translation RMSE (m)0.2858 | 21 | |
| Tracking | TUM-RGBD (various sequences) | Average Translational Error0.035 | 16 | |
| SLAM | TUM RGB-D fr3 office | RMSE (cm)26.802 | 15 | |
| Absolute Pose Estimation | TUM RGB-D v1 | Error (desk)0.016 | 14 | |
| Trajectory Estimation | Replica Static | R Error (0)0.003 | 9 |