Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Omnidata: A Scalable Pipeline for Making Multi-Task Mid-Level Vision Datasets from 3D Scans

About

This paper introduces a pipeline to parametrically sample and render multi-task vision datasets from comprehensive 3D scans from the real world. Changing the sampling parameters allows one to "steer" the generated datasets to emphasize specific information. In addition to enabling interesting lines of research, we show the tooling and generated data suffice to train robust vision models. Common architectures trained on a generated starter dataset reached state-of-the-art performance on multiple common vision tasks and benchmarks, despite having seen no benchmark or non-pipeline data. The depth estimation network outperforms MiDaS and the surface normal estimation network is the first to achieve human-level performance for in-the-wild surface normal estimation -- at least according to one metric on the OASIS benchmark. The Dockerized pipeline with CLI, the (mostly python) code, PyTorch dataloaders for the generated data, the generated starter dataset, download scripts and other utilities are available through our project website, https://omnidata.vision.

Ainaz Eftekhar, Alexander Sax, Roman Bachmann, Jitendra Malik, Amir Zamir• 2021

Related benchmarks

TaskDatasetResultRank
Monocular Depth EstimationKITTI (Eigen)
Abs Rel14.9
502
Depth EstimationNYU v2 (test)
Threshold Accuracy (delta < 1.25)94.5
423
Monocular Depth EstimationNYU v2 (test)
Abs Rel7.4
257
Surface Normal EstimationNYU v2 (test)
Mean Angle Distance (MAD)16.7
206
Monocular Depth EstimationKITTI (Eigen split)
Abs Rel0.149
193
Depth EstimationNYU Depth V2--
177
Monocular Depth EstimationKITTI
Abs Rel14.9
161
Monocular Depth EstimationETH3D
AbsRel16.6
117
Monocular Depth EstimationNYU V2
Delta 1 Acc94.5
113
Monocular Depth EstimationKITTI (test)
Abs Rel Error14.9
103
Showing 10 of 43 rows

Other info

Follow for update