DROID-SLAM: Deep Visual SLAM for Monocular, Stereo, and RGB-D Cameras
About
We introduce DROID-SLAM, a new deep learning based SLAM system. DROID-SLAM consists of recurrent iterative updates of camera pose and pixelwise depth through a Dense Bundle Adjustment layer. DROID-SLAM is accurate, achieving large improvements over prior work, and robust, suffering from substantially fewer catastrophic failures. Despite training on monocular video, it can leverage stereo or RGB-D video to achieve improved performance at test time. The URL to our open source code is https://github.com/princeton-vl/DROID-SLAM.
Zachary Teed, Jia Deng• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Camera pose estimation | Sintel | ATE0.175 | 192 | |
| 3D Reconstruction | 7 Scenes | -- | 94 | |
| Visual-Inertial Odometry | EuRoC (All sequences) | MH1 Error0.013 | 62 | |
| Tracking | TUM RGB-D 44 (various sequences) | Average Error1.62 | 41 | |
| Visual Odometry | TUM-RGBD | freiburg1/desk2 Error0.042 | 37 | |
| Visual Odometry | KITTI | KITTI Seq 03 Error2.38 | 37 | |
| Absolute Trajectory Estimation | TUM RGB-D | Desk Error0.018 | 36 | |
| Pose Estimation | ScanNet++ | -- | 32 | |
| Tracking | TUM 8 dynamic scenes | f3 Walk Scale/Translation Error1.4 | 28 | |
| Camera pose estimation | TUM RGB-D 36 | Error (desk)0.018 | 26 |
Showing 10 of 186 rows
...