DF-Net: Unsupervised Joint Learning of Depth and Flow using Cross-Task Consistency
About
We present an unsupervised learning framework for simultaneously training single-view depth prediction and optical flow estimation models using unlabeled video sequences. Existing unsupervised methods often exploit brightness constancy and spatial smoothness priors to train depth or flow models. In this paper, we propose to leverage geometric consistency as additional supervisory signals. Our core idea is that for rigid regions we can use the predicted scene depth and camera motion to synthesize 2D optical flow by backprojecting the induced 3D scene flow. The discrepancy between the rigid flow (from depth prediction and camera motion) and the estimated flow (from optical flow model) allows us to impose a cross-task consistency loss. While all the networks are jointly optimized during training, they can be applied independently at test time. Extensive experiments demonstrate that our depth and flow models compare favorably with state-of-the-art unsupervised methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Monocular Depth Estimation | KITTI (Eigen) | Abs Rel0.146 | 502 | |
| Optical Flow Estimation | KITTI 2015 (train) | Fl-epe8.98 | 431 | |
| Depth Estimation | KITTI (Eigen split) | RMSE5.507 | 276 | |
| Monocular Depth Estimation | KITTI | Abs Rel0.15 | 161 | |
| Monocular Depth Estimation | KITTI Raw Eigen (test) | RMSE5.215 | 159 | |
| Monocular Depth Estimation | Make3D (test) | Abs Rel0.331 | 132 | |
| Monocular Depth Estimation | KITTI 80m maximum depth (Eigen) | Abs Rel0.146 | 126 | |
| Optical Flow | KITTI 2012 (train) | -- | 115 | |
| Monocular Depth Estimation | KITTI 2015 (Eigen split) | Abs Rel0.15 | 95 | |
| Optical Flow | KITTI 2015 (test) | Fl Error (All)25.7 | 95 |