Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Multi-Space Alignments Towards Universal LiDAR Segmentation

About

A unified and versatile LiDAR segmentation model with strong robustness and generalizability is desirable for safe autonomous driving perception. This work presents M3Net, a one-of-a-kind framework for fulfilling multi-task, multi-dataset, multi-modality LiDAR segmentation in a universal manner using just a single set of parameters. To better exploit data volume and diversity, we first combine large-scale driving datasets acquired by different types of sensors from diverse scenes and then conduct alignments in three spaces, namely data, feature, and label spaces, during the training. As a result, M3Net is capable of taming heterogeneous data for training state-of-the-art LiDAR segmentation models. Extensive experiments on twelve LiDAR segmentation datasets verify our effectiveness. Notably, using a shared set of parameters, M3Net achieves 75.1%, 83.1%, and 72.4% mIoU scores, respectively, on the official benchmarks of SemanticKITTI, nuScenes, and Waymo Open.

Youquan Liu, Lingdong Kong, Xiaoyang Wu, Runnan Chen, Xin Li, Liang Pan, Ziwei Liu, Yuexin Ma• 2024

Related benchmarks

TaskDatasetResultRank
Semantic segmentationSemanticKITTI (test)
mIoU75.1
335
Semantic segmentationnuScenes (val)
mIoU (Segmentation)0.79
212
LiDAR Semantic SegmentationnuScenes (val)
mIoU80.9
169
LiDAR Semantic SegmentationnuScenes official (test)
mIoU83.1
132
LiDAR Semantic SegmentationSemanticKITTI (test)
mIoU75.1
125
Semantic segmentationSemanticKITTI (val)
mIoU72
117
LiDAR Semantic SegmentationSemanticKITTI (val)
mIoU72
87
Semantic segmentationWaymo Open Dataset (val)
mIoU72.4
63
LiDAR-based Panoptic SegmentationnuScenes (val)
PQ71.7
17
Panoptic LiDAR SegmentationPanoptic-SemanticKITTI (val)
PQ63.87
10
Showing 10 of 22 rows

Other info

Code

Follow for update