DepthTransfer: Depth Extraction from Video Using Non-parametric Sampling
About
We describe a technique that automatically generates plausible depth maps from videos using non-parametric depth sampling. We demonstrate our technique in cases where past methods fail (non-translating cameras and dynamic scenes). Our technique is applicable to single images as well as videos. For videos, we use local motion cues to improve the inferred depth maps, while optical flow is used to ensure temporal depth consistency. For training and evaluation, we use a Kinect-based system to collect a large dataset containing stereoscopic videos with known depths. We show that our depth estimation technique outperforms the state-of-the-art on benchmark databases. Our technique can be used to automatically convert a monoscopic video into stereo for 3D visualization, and we demonstrate this through a variety of visually pleasing results for indoor and outdoor scenes, including results from the feature film Charade.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Depth Estimation | NYU v2 (test) | Threshold Accuracy (delta < 1.25)54.2 | 423 | |
| Monocular Depth Estimation | NYU v2 (test) | Abs Rel0.349 | 257 | |
| Depth Estimation | NYU Depth V2 | RMSE1.2 | 177 | |
| Monocular Depth Estimation | Make3D (test) | Abs Rel0.398 | 132 | |
| Depth Prediction | NYU Depth V2 (test) | -- | 113 | |
| Monocular Depth Estimation | NYU Depth Eigen v2 (test) | A.Rel0.349 | 49 | |
| Depth Prediction | Make3D C1 (test) | Log10 Error (log10)0.127 | 27 | |
| Single-view depth estimation | NYUv2 36 (test) | AbsRel0.349 | 21 | |
| Single-view depth estimation | NYU official 654 images v2 (test) | AbsRel0.349 | 21 | |
| Depth Estimation | Make3D (C2 error) | Relative Error (rel)0.361 | 17 |