Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial and Multi-Map SLAM

About

This paper presents ORB-SLAM3, the first system able to perform visual, visual-inertial and multi-map SLAM with monocular, stereo and RGB-D cameras, using pin-hole and fisheye lens models. The first main novelty is a feature-based tightly-integrated visual-inertial SLAM system that fully relies on Maximum-a-Posteriori (MAP) estimation, even during the IMU initialization phase. The result is a system that operates robustly in real-time, in small and large, indoor and outdoor environments, and is 2 to 5 times more accurate than previous approaches. The second main novelty is a multiple map system that relies on a new place recognition method with improved recall. Thanks to it, ORB-SLAM3 is able to survive to long periods of poor visual information: when it gets lost, it starts a new map that will be seamlessly merged with previous maps when revisiting mapped areas. Compared with visual odometry systems that only use information from the last few seconds, ORB-SLAM3 is the first system able to reuse in all the algorithm stages all previous information. This allows to include in bundle adjustment co-visible keyframes, that provide high parallax observations boosting accuracy, even if they are widely separated in time or if they come from a previous mapping session. Our experiments show that, in all sensor configurations, ORB-SLAM3 is as robust as the best systems available in the literature, and significantly more accurate. Notably, our stereo-inertial SLAM achieves an average accuracy of 3.6 cm on the EuRoC drone and 9 mm under quick hand-held motions in the room of TUM-VI dataset, a setting representative of AR/VR scenarios. For the benefit of the community we make public the source code.

Carlos Campos, Richard Elvira, Juan J. G\'omez Rodr\'iguez, Jos\'e M. M. Montiel, Juan D. Tard\'os• 2020

Related benchmarks

TaskDatasetResultRank
OdometryKITTI 2011 (sequences 0036, 0101, 0028, 0034, 0042, 0022, 0071, 0018, 0020, 0027, 0033, 0027, 0034, 0047)
RTE3.1
60
Visual-Inertial OdometryEuRoC (All sequences)
MH1 Error0.016
51
TrackingStrided EuRoC
MH 01 Sequence Result0.00e+0
48
Visual OdometryTUM-RGBD
freiburg1/xyz Error0.009
34
TrackingAriaMultiagent--
30
TrackingTUM RGB-D 44 (various sequences)
Average Error12.62
28
Visual OdometryKITTI
KITTI Seq 03 Error1.036
27
Camera TrackingBONN dynamic sequences--
25
SLAMROTIO
Success Rate (%)1
24
Absolute Trajectory EstimationTUM RGB-D
Desk Error0.017
23
Showing 10 of 133 rows
...

Other info

Follow for update