GeoNet: Unsupervised Learning of Dense Depth, Optical Flow and Camera Pose
About
We propose GeoNet, a jointly unsupervised learning framework for monocular depth, optical flow and ego-motion estimation from videos. The three components are coupled by the nature of 3D scene geometry, jointly learned by our framework in an end-to-end manner. Specifically, geometric relationships are extracted over the predictions of individual modules and then combined as an image reconstruction loss, reasoning about static and dynamic scene parts separately. Furthermore, we propose an adaptive geometric consistency loss to increase robustness towards outliers and non-Lambertian regions, which resolves occlusions and texture ambiguities effectively. Experimentation on the KITTI driving dataset reveals that our scheme achieves state-of-the-art results in all of the three tasks, performing better than previously unsupervised methods and comparably with supervised ones.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Monocular Depth Estimation | KITTI (Eigen) | Abs Rel0.147 | 502 | |
| Optical Flow Estimation | KITTI 2015 (train) | Fl-epe10.81 | 431 | |
| Depth Estimation | KITTI (Eigen split) | RMSE5.567 | 276 | |
| Surface Normal Estimation | NYU v2 (test) | Mean Angle Distance (MAD)19 | 206 | |
| Monocular Depth Estimation | KITTI | Abs Rel0.155 | 161 | |
| Monocular Depth Estimation | KITTI Raw Eigen (test) | RMSE5.567 | 159 | |
| Monocular Depth Estimation | KITTI 80m maximum depth (Eigen) | Abs Rel0.153 | 126 | |
| Monocular Depth Estimation | KITTI 2015 (Eigen split) | Abs Rel0.147 | 95 | |
| Optical Flow | KITTI 2015 (test) | -- | 95 | |
| Depth Prediction | KITTI original ground truth (test) | Abs Rel0.155 | 38 |