Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

g3D-LF: Generalizable 3D-Language Feature Fields for Embodied Tasks

About

We introduce Generalizable 3D-Language Feature Fields (g3D-LF), a 3D representation model pre-trained on large-scale 3D-language dataset for embodied tasks. Our g3D-LF processes posed RGB-D images from agents to encode feature fields for: 1) Novel view representation predictions from any position in the 3D scene; 2) Generations of BEV maps centered on the agent; 3) Querying targets using multi-granularity language within the above-mentioned representations. Our representation can be generalized to unseen environments, enabling real-time construction and dynamic updates. By volume rendering latent features along sampled rays and integrating semantic and spatial relationships through multiscale encoders, our g3D-LF produces representations at different scales and perspectives, aligned with multi-granularity language, via multi-level contrastive learning. Furthermore, we prepare a large-scale 3D-language dataset to align the representations of the feature fields with language. Extensive experiments on Vision-and-Language Navigation under both Panorama and Monocular settings, Zero-shot Object Navigation, and Situated Question Answering tasks highlight the significant advantages and effectiveness of our g3D-LF for embodied tasks.

Zihan Wang, Gim Hee Lee• 2024

Related benchmarks

TaskDatasetResultRank
Vision-Language NavigationR2R-CE (val-unseen)
Success Rate (SR)61
266
Vision-and-Language NavigationREVERIE (val unseen)
SPL23.8
129
Vision-and-Language NavigationR2R-CE (test-unseen)
SR58
50
Vision-and-Language NavigationR2R-CE v1.0 (val unseen)
NE (Navigation Error)4.53
19
Embodied NavigationR2R-CE
Navigation Error (NE)5.7
19
Vision-and-Language NavigationREVERIE CE (val unseen)
NE6.5
8
Embodied NavigationNavRAG-CE
Navigation Error (NE)8.85
5
Showing 7 of 7 rows

Other info

Follow for update