Asymmetric Dual Self-Distillation for 3D Self-Supervised Representation Learning
About
Learning semantically meaningful representations from unstructured 3D point clouds remains a central challenge in computer vision, especially in the absence of large-scale labeled datasets. While masked point modeling (MPM) is widely used in self-supervised 3D learning, its reconstruction-based objective can limit its ability to capture high-level semantics. We propose AsymDSD, an Asymmetric Dual Self-Distillation framework that unifies masked modeling and invariance learning through prediction in the latent space rather than the input space. AsymDSD builds on a joint embedding architecture and introduces several key design choices: an efficient asymmetric setup, disabling attention between masked queries to prevent shape leakage, multi-mask sampling, and a point cloud adaptation of multi-crop. AsymDSD achieves state-of-the-art results on ScanObjectNN (90.53%) and further improves to 93.72% when pretrained on 930k shapes, surpassing prior methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Part Segmentation | ShapeNetPart (test) | mIoU (Inst.)86 | 312 | |
| Object Classification | ScanObjectNN OBJ_BG | Accuracy97.07 | 215 | |
| Object Classification | ScanObjectNN PB_T50_RS | Accuracy93.72 | 195 | |
| Object Classification | ScanObjectNN OBJ_ONLY | Overall Accuracy94.83 | 166 | |
| 3D Point Cloud Classification | ModelNet40 | Accuracy94.7 | 69 | |
| 3D Object Classification | ModelNet40 few-shot | Accuracy98.8 | 60 |