Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Point Transformer

About

Self-attention networks have revolutionized natural language processing and are making impressive strides in image analysis tasks such as image classification and object detection. Inspired by this success, we investigate the application of self-attention networks to 3D point cloud processing. We design self-attention layers for point clouds and use these to construct self-attention networks for tasks such as semantic scene segmentation, object part segmentation, and object classification. Our Point Transformer design improves upon prior work across domains and tasks. For example, on the challenging S3DIS dataset for large-scale semantic scene segmentation, the Point Transformer attains an mIoU of 70.4% on Area 5, outperforming the strongest prior model by 3.3 absolute percentage points and crossing the 70% mIoU threshold for the first time.

Hengshuang Zhao, Li Jiang, Jiaya Jia, Philip Torr, Vladlen Koltun• 2020

Related benchmarks

TaskDatasetResultRank
Semantic segmentationS3DIS (Area 5)
mIOU70.6
907
Semantic segmentationS3DIS (6-fold)
mIoU (Mean IoU)73.5
344
Semantic segmentationScanNet V2 (val)
mIoU70.6
316
Part SegmentationShapeNetPart (test)
mIoU (Inst.)86.6
312
3D Object ClassificationModelNet40 (test)
Accuracy92.8
308
3D Point Cloud ClassificationModelNet40 (test)
OA93.7
297
Semantic segmentationScanNet (val)
mIoU70.6
274
Shape classificationModelNet40 (test)
OA93.7
255
Semantic segmentationScanNet v2 (test)
mIoU70.6
248
Part SegmentationShapeNetPart
mIoU (Instance)86.6
246
Showing 10 of 73 rows
...

Other info

Follow for update