Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SAL: Sign Agnostic Learning of Shapes from Raw Data

About

Recently, neural networks have been used as implicit representations for surface reconstruction, modelling, learning, and generation. So far, training neural networks to be implicit representations of surfaces required training data sampled from a ground-truth signed implicit functions such as signed distance or occupancy functions, which are notoriously hard to compute. In this paper we introduce Sign Agnostic Learning (SAL), a deep learning approach for learning implicit shape representations directly from raw, unsigned geometric data, such as point clouds and triangle soups. We have tested SAL on the challenging problem of surface reconstruction from an un-oriented point cloud, as well as end-to-end human shape space learning directly from raw scans dataset, and achieved state of the art reconstructions compared to current approaches. We believe SAL opens the door to many geometric deep learning applications with real-world data, alleviating the usual painstaking, often manual pre-process.

Matan Atzmon, Yaron Lipman• 2019

Related benchmarks

TaskDatasetResultRank
Surface ReconstructionSurface Reconstruction Benchmark (SRB) 5 noisy range scans
Dist Error (c) vs GT0.36
15
Point Cloud CompletionShapeNet unprocessed cars
Chamfer Distance L26.39e-4
12
Surface ReconstructionShapeNet 260 shapes 15
sCD (mean)0.0011
9
Distance QueryShapeNet
RMSE (mean)0.0251
7
Surface ReconstructionShapeNet-55 (test)
mIoU74
7
Surface ReconstructionSurface Reconstruction Benchmark
Chamfer Distance (d_C)0.36
6
Surface ReconstructionSRB GT
Anchor dc0.42
6
Surface ReconstructionSRB Scans
Anchor DC Error0.17
6
Surface ReconstructionShapeNet
Squared Chamfer Distance (Mean)0.0011
6
Showing 9 of 9 rows

Other info

Follow for update