Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

GeoAuxNet: Towards Universal 3D Representation Learning for Multi-sensor Point Clouds

About

Point clouds captured by different sensors such as RGB-D cameras and LiDAR possess non-negligible domain gaps. Most existing methods design different network architectures and train separately on point clouds from various sensors. Typically, point-based methods achieve outstanding performances on even-distributed dense point clouds from RGB-D cameras, while voxel-based methods are more efficient for large-range sparse LiDAR point clouds. In this paper, we propose geometry-to-voxel auxiliary learning to enable voxel representations to access point-level geometric information, which supports better generalisation of the voxel-based backbone with additional interpretations of multi-sensor point clouds. Specifically, we construct hierarchical geometry pools generated by a voxel-guided dynamic point network, which efficiently provide auxiliary fine-grained geometric information adapted to different stages of voxel features. We conduct experiments on joint multi-sensor datasets to demonstrate the effectiveness of GeoAuxNet. Enjoying elaborate geometric information, our method outperforms other models collectively trained on multi-sensor datasets, and achieve competitive results with the-state-of-art experts on each single dataset.

Shengjun Zhang, Xin Fei, Yueqi Duan• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationScanNet (val)
mIoU71.3
231
Semantic segmentationSemanticKITTI (val)
mIoU63.8
117
3D Semantic SegmentationS3DIS Area 5 (test)
mIoU (%)69.5
32
Showing 3 of 3 rows

Other info

Code

Follow for update