3DFeat-Net: Weakly Supervised Local 3D Features for Point Cloud Registration
About
In this paper, we propose the 3DFeat-Net which learns both 3D feature detector and descriptor for point cloud matching using weak supervision. Unlike many existing works, we do not require manual annotation of matching point clusters. Instead, we leverage on alignment and attention mechanisms to learn feature correspondences from GPS/INS tagged 3D point clouds without explicitly specifying them. We create training and benchmark outdoor Lidar datasets, and experiments show that 3DFeat-Net obtains state-of-the-art performance on these gravity-aligned datasets.
Zi Jian Yew, Gim Hee Lee• 2018
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Point cloud registration | KITTI odometry (sequences 8-10) | Success Rate96 | 70 | |
| Point cloud registration | ETH | Success Rate100 | 38 | |
| Point cloud registration | Oxford Stereo DSO | RTE (m)0.38 | 30 | |
| Geometric Registration | KITTI | RTE0.246 | 16 | |
| 3D local descriptor matching | 3DMatch | -- | 16 | |
| Geometric Registration | KITTI Dataset (test) | RTE0.246 | 14 | |
| Geometric Registration | Oxford Dataset (test) | RTE0.3 | 13 | |
| Geometric Registration | KITTI odometry dataset | Success Rate95.97 | 7 | |
| Point cloud registration | KITTI odometry | RTE Avg (cm)25.9 | 5 |
Showing 9 of 9 rows