Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

VLM-Grounder: A VLM Agent for Zero-Shot 3D Visual Grounding

About

3D visual grounding is crucial for robots, requiring integration of natural language and 3D scene understanding. Traditional methods depending on supervised learning with 3D point clouds are limited by scarce datasets. Recently zero-shot methods leveraging LLMs have been proposed to address the data issue. While effective, these methods only use object-centric information, limiting their ability to handle complex queries. In this work, we present VLM-Grounder, a novel framework using vision-language models (VLMs) for zero-shot 3D visual grounding based solely on 2D images. VLM-Grounder dynamically stitches image sequences, employs a grounding and feedback scheme to find the target object, and uses a multi-view ensemble projection to accurately estimate 3D bounding boxes. Experiments on ScanRefer and Nr3D datasets show VLM-Grounder outperforms previous zero-shot methods, achieving 51.6% Acc@0.25 on ScanRefer and 48.0% Acc on Nr3D, without relying on 3D geometry or object priors. Codes are available at https://github.com/OpenRobotLab/VLM-Grounder .

Runsen Xu, Zhiwei Huang, Tai Wang, Yilun Chen, Jiangmiao Pang, Dahua Lin• 2024

Related benchmarks

TaskDatasetResultRank
3D Visual GroundingNr3D (test)
Overall Success Rate48
88
3D Visual GroundingNr3D
Overall Success Rate48
74
3D Visual GroundingScanRefer Unique
Acc@0.25 (IoU=0.25)51.6
24
3D Visual GroundingScanRefer
Acc@0.2566
23
3D Visual GroundingScanRefer Overall
Acc @ 0.2548.3
17
3D Visual GroundingScanRefer 250 scenes (test)
Acc@0.25 (Unique)66
7
3D Visual GroundingOpenTarget randomly selected 300 samples
Accuracy @ IoU=0.2528.6
6
Showing 7 of 7 rows

Other info

Follow for update