Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Closer Look at Invariances in Self-supervised Pre-training for 3D Vision

About

Self-supervised pre-training for 3D vision has drawn increasing research interest in recent years. In order to learn informative representations, a lot of previous works exploit invariances of 3D features, e.g., perspective-invariance between views of the same scene, modality-invariance between depth and RGB images, format-invariance between point clouds and voxels. Although they have achieved promising results, previous researches lack a systematic and fair comparison of these invariances. To address this issue, our work, for the first time, introduces a unified framework, under which various pre-training methods can be investigated. We conduct extensive experiments and provide a closer look at the contributions of different invariances in 3D pre-training. Also, we propose a simple but effective method that jointly pre-trains a 3D encoder and a depth map encoder using contrastive learning. Models pre-trained with our method gain significant performance boost in downstream tasks. For instance, a pre-trained VoteNet outperforms previous methods on SUN RGB-D and ScanNet object detection benchmarks with a clear margin.

Lanxiao Li, Michael Heizmann• 2022

Related benchmarks

TaskDatasetResultRank
3D Object DetectionScanNet V2 (val)
mAP@0.2564.2
352
3D Object DetectionSUN RGB-D
mAP@0.2560.2
104
3D Object DetectionSUN RGB-D v1 (val)
mAP@0.2560.2
81
3D Object DetectionScanNet V2
AP5041.5
54
Showing 4 of 4 rows

Other info

Follow for update