Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Depth Field Networks for Generalizable Multi-view Scene Representation

About

Modern 3D computer vision leverages learning to boost geometric reasoning, mapping image data to classical structures such as cost volumes or epipolar constraints to improve matching. These architectures are specialized according to the particular problem, and thus require significant task-specific tuning, often leading to poor domain generalization performance. Recently, generalist Transformer architectures have achieved impressive results in tasks such as optical flow and depth estimation by encoding geometric priors as inputs rather than as enforced constraints. In this paper, we extend this idea and propose to learn an implicit, multi-view consistent scene representation, introducing a series of 3D data augmentation techniques as a geometric inductive prior to increase view diversity. We also show that introducing view synthesis as an auxiliary task further improves depth estimation. Our Depth Field Networks (DeFiNe) achieve state-of-the-art results in stereo and video depth estimation without explicit geometric constraints, and improve on zero-shot domain generalization by a wide margin.

Vitor Guizilini, Igor Vasiljevic, Jiading Fang, Rares Ambrus, Greg Shakhnarovich, Matthew Walter, Adrien Gaidon• 2022

Related benchmarks

TaskDatasetResultRank
Depth EstimationScanNet (test)
Abs Rel0.056
65
Depth EstimationSUN3D
Abs Rel0.095
13
Depth EstimationRGBD
Abs Rel0.095
10
Video Depth EstimationScanNet (in-domain)
Abs Rel0.059
8
Stereo Depth EstimationScanNet
Abs. Rel. Error0.093
7
Showing 5 of 5 rows

Other info

Follow for update