Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Unsupervised Depth Completion from Visual Inertial Odometry

About

We describe a method to infer dense depth from camera motion and sparse depth as estimated using a visual-inertial odometry system. Unlike other scenarios using point clouds from lidar or structured light sensors, we have few hundreds to few thousand points, insufficient to inform the topology of the scene. Our method first constructs a piecewise planar scaffolding of the scene, and then uses it to infer dense depth using the image along with the sparse points. We use a predictive cross-modal criterion, akin to `self-supervision,' measuring photometric consistency across time, forward-backward pose consistency, and geometric compatibility with the sparse point cloud. We also launch the first visual-inertial + depth dataset, which we hope will foster additional exploration into combining the complementary strengths of visual and inertial sensors. To compare our method to prior work, we adopt the unsupervised KITTI depth completion benchmark, and show state-of-the-art performance on it. Code available at: https://github.com/alexklwong/unsupervised-depth-completion-visual-inertial-odometry.

Alex Wong, Xiaohan Fei, Stephanie Tsuei, Stefano Soatto• 2019

Related benchmarks

TaskDatasetResultRank
Depth CompletionNYU-depth-v2 official (test)
RMSE228.4
187
Depth CompletionKITTI depth completion official (test)
RMSE (mm)1.17e+3
154
Depth CompletionKITTI (test)
RMSE1.17e+3
67
Depth CompletionKITTI online leaderboard (test)
MAE299.4
48
Depth CompletionNYU Depth V2
RMSE0.228
34
Depth CompletionKITTI depth completion (val)
RMSE (mm)1.23e+3
34
Depth CompletionKITTI-Depth
MAE299.4
27
Depth CompletionVOID (test)
MAE73.14
18
Depth CompletionVOID 1.0 (test)
MAE73.14
7
Depth Completion (Varying Density)VOID 1.0 (test)
MAE73.14
6
Showing 10 of 10 rows

Other info

Code

Follow for update