DeMoN: Depth and Motion Network for Learning Monocular Stereo
About
In this paper we formulate structure from motion as a learning problem. We train a convolutional network end-to-end to compute depth and camera motion from successive, unconstrained image pairs. The architecture is composed of multiple stacked encoder-decoder networks, the core part being an iterative network that is able to improve its own predictions. The network estimates not only depth and motion, but additionally surface normals, optical flow between the images and confidence of the matching. A crucial component of the approach is a training loss based on spatial relative differences. Compared to traditional two-frame structure from motion methods, results are more accurate and more robust. In contrast to the popular depth-from-single-image networks, DeMoN learns the concept of matching and, thus, better generalizes to structures not seen during training.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Depth Estimation | ScanNet (test) | Abs Rel0.231 | 65 | |
| Depth Estimation | Sun3D (test) | Abs Rel17.2 | 22 | |
| 2D Depth Estimation | 7 Scenes | Abs Rel0.3888 | 20 | |
| Depth Estimation | Scenes11 (test) | L1 Relative Error0.248 | 12 | |
| Depth Estimation | RGBD-SLAM (test) | Abs Rel0.1569 | 10 | |
| Video Depth Estimation | ScanNet++ | Absolute Relative Error75 | 10 | |
| Pose Estimation | Scenes11 (test) | Rotation Error0.809 | 8 | |
| Pose Estimation | MVS DeMoN version (test) | Rot Error5.156 | 8 | |
| Pose Estimation | Sun3D (test) | Rotation Error1.801 | 8 | |
| Two-view Depth Estimation | ScanNet (test) | Abs Rel0.231 | 8 |