PPF-FoldNet: Unsupervised Learning of Rotation Invariant 3D Local Descriptors
About
We present PPF-FoldNet for unsupervised learning of 3D local descriptors on pure point cloud geometry. Based on the folding-based auto-encoding of well known point pair features, PPF-FoldNet offers many desirable properties: it necessitates neither supervision, nor a sensitive local reference frame, benefits from point-set sparsity, is end-to-end, fast, and can extract powerful rotation invariant descriptors. Thanks to a novel feature visualization, its evolution can be monitored to provide interpretable insights. Our extensive experiments demonstrate that despite having six degree-of-freedom invariance and lack of training labels, our network achieves state of the art results in standard benchmark datasets and outperforms its competitors when rotations and varying point densities are present. PPF-FoldNet achieves $9\%$ higher recall on standard benchmarks, $23\%$ higher recall when rotations are introduced into the same datasets and finally, a margin of $>35\%$ is attained when point density is significantly decreased.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Point cloud registration | 3DMatch (test) | -- | 339 | |
| Point cloud registration | 3DLoMatch Indoor (test) | RR24.8 | 66 | |
| Feature Matching | 3DMatch (Origin) | STD10.5 | 33 | |
| Local Descriptor Matching | 3DMatch 1.0 (test) | Kitchen Scene Performance78.85 | 18 | |
| Descriptor matching | 3DMatch Rotated | STD10.4 | 18 | |
| 3D local descriptor matching | 3DMatch | Average Recall71.8 | 16 | |
| Feature Matching | 3DMatch Rotated (test) | FMR0.731 | 12 | |
| Point cloud registration | 3DMatch Indoor (test) | Rotation Rate69.3 | 11 | |
| 3D Fragment Matching | 3DMatch (test) | Kitchen Success Rate78.66 | 9 | |
| 3D Keypoint Matching | 3DMatch Rotated (test) | AR (tau2=0.05)73.1 | 9 |