DeepFactors: Real-Time Probabilistic Dense Monocular SLAM
About
The ability to estimate rich geometry and camera motion from monocular imagery is fundamental to future interactive robotics and augmented reality applications. Different approaches have been proposed that vary in scene geometry representation (sparse landmarks, dense maps), the consistency metric used for optimising the multi-view problem, and the use of learned priors. We present a SLAM system that unifies these methods in a probabilistic framework while still maintaining real-time performance. This is achieved through the use of a learned compact depth map representation and reformulating three different types of errors: photometric, reprojection and geometric, which we make use of within standard factor graph software. We evaluate our system on trajectory estimation and depth reconstruction on real-world sequences and present various examples of estimated dense geometry.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Visual-Inertial Odometry | EuRoC (All sequences) | MH1 Error1.587 | 51 | |
| Camera pose estimation | TUM freiburg1 | Rotation Error0.043 | 34 | |
| Visual Odometry | TUM-RGBD | freiburg1/xyz Error0.035 | 34 | |
| Absolute Trajectory Estimation | TUM RGB-D | Desk Error0.17 | 23 | |
| Tracking | TUM-RGBD (various sequences) | Average Translational Error0.233 | 16 | |
| Absolute Pose Estimation | TUM RGB-D v1 | Error (desk)0.17 | 14 | |
| Monocular SLAM | EuRoC (test) | ATE Error (MH03)3.139 | 12 | |
| Simultaneous Localization and Mapping (SLAM) | TUM-RGBD (various sequences) | Error Desk0.17 | 8 |