Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Representation Learning for Point Cloud Understanding

About

With the rapid advancement of technology, 3D data acquisition and utilization have become increasingly prevalent across various fields, including computer vision, robotics, and geospatial analysis. 3D data, captured through methods such as 3D scanners, LiDARs, and RGB-D cameras, provides rich geometric, shape, and scale information. When combined with 2D images, 3D data offers machines a comprehensive understanding of their environment, benefiting applications like autonomous driving, robotics, remote sensing, and medical treatment. This dissertation focuses on three main areas: supervised representation learning for point cloud primitive segmentation, self-supervised learning methods, and transfer learning from 2D to 3D. Our approach, which integrates pre-trained 2D models to support 3D network training, significantly improves 3D understanding without merely transforming 2D data. Extensive experiments validate the effectiveness of our methods, showcasing their potential to advance point cloud representation learning by effectively integrating 2D knowledge.

Siming Yan• 2025

Related benchmarks

TaskDatasetResultRank
Semantic segmentationS3DIS (Area 5)
mIOU73.8
799
3D Object DetectionScanNet V2 (val)
mAP@0.2576.1
352
Semantic segmentationS3DIS (6-fold)
mIoU (Mean IoU)78.3
315
Shape classificationModelNet40 (test)--
255
Object ClassificationScanObjectNN OBJ_BG
Accuracy92.5
215
3D Object DetectionScanNet
mAP@0.2576.1
123
3D Object Part SegmentationShapeNet Part (test)--
114
3D Object DetectionSUN RGB-D
mAP@0.2567.6
104
3D Object DetectionSUN RGB-D v1 (val)
mAP@0.2567.6
81
3D Object ClassificationScanObjectNN PB_T50_RS
OA91
72
Showing 10 of 23 rows

Other info

Follow for update