SL3D: Self-supervised-Self-labeled 3D Recognition
About
Deep learning has attained remarkable success in many 3D visual recognition tasks, including shape classification, object detection, and semantic segmentation. However, many of these results rely on manually collecting densely annotated real-world 3D data, which is highly time-consuming and expensive to obtain, limiting the scalability of 3D recognition tasks. Thus, we study unsupervised 3D recognition and propose a Self-supervised-Self-Labeled 3D Recognition (SL3D) framework. SL3D simultaneously solves two coupled objectives, i.e., clustering and learning feature representation to generate pseudo-labeled data for unsupervised 3D recognition. SL3D is a generic framework and can be applied to solve different 3D recognition tasks, including classification, object detection, and semantic segmentation. Extensive experiments demonstrate its effectiveness. Code is available at https://github.com/fcendra/sl3d.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| 3D Object Classification | ModelNet40 (test) | Accuracy77.2 | 302 | |
| 3D Semantic Segmentation | ScanNet V2 (val) | mIoU32.9 | 171 | |
| 3D Semantic Segmentation | ScanNet v2 (test) | mIoU1.05e+3 | 110 | |
| 3D Object Detection | ScanNet (val) | mAP@0.259.3 | 66 | |
| 3D Semantic Segmentation | ScanNet v2 (train) | mIoU60.2 | 8 | |
| 3D Object Detection | ScanNet SL3D (val) | mAP@0.2520.3 | 4 | |
| 3D Semantic Segmentation | ScanNet sub-cloud v2 (val) | -- | 2 |