Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

360SD-Net: 360{\deg} Stereo Depth Estimation with Learnable Cost Volume

About

Recently, end-to-end trainable deep neural networks have significantly improved stereo depth estimation for perspective images. However, 360{\deg} images captured under equirectangular projection cannot benefit from directly adopting existing methods due to distortion introduced (i.e., lines in 3D are not projected onto lines in 2D). To tackle this issue, we present a novel architecture specifically designed for spherical disparity using the setting of top-bottom 360{\deg} camera pairs. Moreover, we propose to mitigate the distortion issue by (1) an additional input branch capturing the position and relation of each pixel in the spherical coordinate, and (2) a cost volume built upon a learnable shifting filter. Due to the lack of 360{\deg} stereo data, we collect two 360{\deg} stereo datasets from Matterport3D and Stanford3D for training and evaluation. Extensive experiments and ablation study are provided to validate our method against existing algorithms. Finally, we show promising results on real-world environments capturing images with two consumer-level cameras.

Ning-Hsu Wang, Bolivar Solarte, Yi-Hsuan Tsai, Wei-Chen Chiu, Min Sun• 2019

Related benchmarks

TaskDatasetResultRank
Stereo Depth EstimationMP3D
Disparity MAE0.1447
6
Stereo Depth EstimationSF3D
Disparity MAE0.1034
6
Omnidirectional Stereo Disparity EstimationHelvipad (test)
MAE0.224
5
Omnidirectional Stereo Depth EstimationHelvipad (test)
MAE2.122
5
Showing 4 of 4 rows

Other info

Code

Follow for update