Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DeFeat-Net: General Monocular Depth via Simultaneous Unsupervised Representation Learning

About

In the current monocular depth research, the dominant approach is to employ unsupervised training on large datasets, driven by warped photometric consistency. Such approaches lack robustness and are unable to generalize to challenging domains such as nighttime scenes or adverse weather conditions where assumptions about photometric consistency break down. We propose DeFeat-Net (Depth & Feature network), an approach to simultaneously learn a cross-domain dense feature representation, alongside a robust depth-estimation framework based on warped feature consistency. The resulting feature representation is learned in an unsupervised manner with no explicit ground-truth correspondences required. We show that within a single domain, our technique is comparable to both the current state of the art in monocular depth estimation and supervised feature representation learning. However, by simultaneously learning features, depth and motion, our technique is able to generalize to challenging domains, allowing DeFeat-Net to outperform the current state-of-the-art with around 10% reduction in all error measures on more challenging sequences such as nighttime driving.

Jaime Spencer, Richard Bowden, Simon Hadfield• 2020

Related benchmarks

TaskDatasetResultRank
Monocular Depth EstimationKITTI (Eigen)
Abs Rel0.126
502
Stereo MatchingKITTI 2015 (test)--
144
Stereo MatchingMiddlebury (test)
3PE6.95
47
Stereo MatchingInria SLFD
3 Pixel Error10.47
41
Stereo MatchingHCI (test)
3PE5.13
35
Stereo MatchingMiddlebury--
34
Monocular Depth EstimationSCARED
Abs Rel0.077
24
Depth EstimationSCARED (test)
Abs Rel0.077
21
Stereo MatchingHCI
3PE12.13
6
Showing 9 of 9 rows

Other info

Follow for update