Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

LatentBKI: Open-Dictionary Continuous Mapping in Visual-Language Latent Spaces with Quantifiable Uncertainty

About

This paper introduces a novel probabilistic mapping algorithm, LatentBKI, which enables open-vocabulary mapping with quantifiable uncertainty. Traditionally, semantic mapping algorithms focus on a fixed set of semantic categories which limits their applicability for complex robotic tasks. Vision-Language (VL) models have recently emerged as a technique to jointly model language and visual features in a latent space, enabling semantic recognition beyond a predefined, fixed set of semantic classes. LatentBKI recurrently incorporates neural embeddings from VL models into a voxel map with quantifiable uncertainty, leveraging the spatial correlations of nearby observations through Bayesian Kernel Inference (BKI). LatentBKI is evaluated against similar explicit semantic mapping and VL mapping frameworks on the popular Matterport3D and Semantic KITTI datasets, demonstrating that LatentBKI maintains the probabilistic benefits of continuous mapping with the additional benefit of open-dictionary queries. Real-world experiments demonstrate applicability to challenging indoor environments.

Joey Wilson, Ruihan Xu, Yile Sun, Parker Ewen, Minghan Zhu, Kira Barton, Maani Ghaffari• 2024

Related benchmarks

TaskDatasetResultRank
Semantic MappingSemanticKITTI (scenes 6-10)
FPS0.53
6
Semantic MappingSceneNet Closed-set (12 full scenes)
FPS1.1
6
Semantic MappingSceneNet (SN)
mIoU4.9
5
Geometric Mapping AccuracyKITTI
Avg Chamfer Distance4
4
Semantic MappingKITTI
mIoU16.5
4
Geometric Mapping AccuracySceneNet (SN)
Avg Chamfer Distance1.189
4
Semantic MappingSceneNet Open-set (12 full scenes)
FPS1.67
2
Showing 7 of 7 rows

Other info

Follow for update