Cross-view Domain Generalization via Geometric Consistency for LiDAR Semantic Segmentation
About
Domain-generalized LiDAR semantic segmentation (LSS) seeks to train models on source-domain point clouds that generalize reliably to multiple unseen target domains, which is essential for real-world LiDAR applications. However, existing approaches assume similar acquisition views (e.g., vehicle-mounted) and struggle in cross-view scenarios, where observations differ substantially due to viewpoint-dependent structural incompleteness and non-uniform point density. Accordingly, we formulate cross-view domain generalization for LiDAR semantic segmentation and propose a novel framework, termed CVGC (Cross-View Geometric Consistency). Specifically, we introduce a cross-view geometric augmentation module that models viewpoint-induced variations in visibility and sampling density, generating multiple cross-view observations of the same scene. Subsequently, a geometric consistency module enforces consistent semantic and occupancy predictions across geometrically augmented point clouds of the same scene. Extensive experiments on six public LiDAR datasets establish the first systematic evaluation of cross-view domain generalization for LiDAR semantic segmentation, demonstrating that CVGC consistently outperforms state-of-the-art methods when generalizing from a single source domain to multiple target domains with heterogeneous acquisition viewpoints. The source code will be publicly available at https://github.com/KintomZi/CVGC-DG
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| LiDAR Semantic Segmentation | ISPRS Vaihingen | Ground IoU0.8669 | 6 | |
| Semantic segmentation | STPLS3D to Toronto-3D (test) | Ground IoU96.35 | 6 | |
| Semantic segmentation | STPLS3D to DALES (test) | Ground IoU94.25 | 6 | |
| LiDAR Semantic Segmentation | Paris-Lille-3D | Ground IoU92.74 | 6 |