Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Multi3DRefer: Grounding Text Description to Multiple 3D Objects

About

We introduce the task of localizing a flexible number of objects in real-world 3D scenes using natural language descriptions. Existing 3D visual grounding tasks focus on localizing a unique object given a text description. However, such a strict setting is unnatural as localizing potentially multiple objects is a common need in real-world scenarios and robotic tasks (e.g., visual navigation and object rearrangement). To address this setting we propose Multi3DRefer, generalizing the ScanRefer dataset and task. Our dataset contains 61926 descriptions of 11609 objects, where zero, single or multiple target objects are referenced by each description. We also introduce a new evaluation metric and benchmark methods from prior work to enable further investigation of multi-modal 3D scene understanding. Furthermore, we develop a better baseline leveraging 2D features from CLIP by rendering object proposals online with contrastive learning, which outperforms the state of the art on the ScanRefer benchmark.

Yiming Zhang, ZeMing Gong, Angel X. Chang• 2023

Related benchmarks

TaskDatasetResultRank
3D Visual GroundingScanRefer (val)
Overall Accuracy @ IoU 0.5044.7
192
3D Visual GroundingScanRefer
Acc@0.544.7
142
3D Dense CaptioningScan2Cap
CIDEr @0.538.4
96
3D Visual GroundingNr3D (test)
Overall Success Rate49.4
88
Referring 3D Instance SegmentationScanRefer (val)
mIoU35.7
37
Visual GroundingScanRefer v1 (val)
Acc@0.5 (All)45.7
30
Multi-object 3D Visual GroundingMulti3DRefer
F1@0.2542.8
24
3D Visual GroundingScanRefer (test)
Unique Accuracy77.2
21
3D Visual GroundingScanRefer v1 (test)
Unique Acc@0.5IoU70.9
15
3D Visual GroundingMulti3DRefer (val)
F1@0.5038.4
14
Showing 10 of 21 rows

Other info

Follow for update