Global Commander and Local Operative: A Dual-Agent Framework for Scene Navigation
About
Vision-and-Language Scene navigation is a fundamental capability for embodied human-AI collaboration, requiring agents to follow natural language instructions to execute coherent action sequences in complex environments. Existing approaches either rely on multiple agents, incurring high coordination and resource costs, or adopt a single-agent paradigm, which overloads the agent with both global planning and local perception, often leading to degraded reasoning and instruction drift in long-horizon settings. To address these issues, we introduce DACo, a planning-grounding decoupled architecture that disentangles global deliberation from local grounding. Concretely, it employs a Global Commander for high-level strategic planning and a Local Operative for egocentric observing and fine-grained execution. By disentangling global reasoning from local action, DACo alleviates cognitive overload and improves long-horizon stability. The framework further integrates dynamic subgoal planning and adaptive replanning to enable structured and resilient navigation. Extensive evaluations on R2R, REVERIE, and R4R demonstrate that DACo achieves 4.9%, 6.5%, 5.4% absolute improvements over the best-performing baselines in zero-shot settings, and generalizes effectively across both closed-source (e.g., GPT-4o) and open-source (e.g., Qwen-VL Series) backbones. DACo provides a principled and extensible paradigm for robust long-horizon navigation. Project page: https://github.com/ChocoWu/DACo
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Vision-and-Language Navigation | R2R (val unseen) | Success Rate (SR)48 | 260 | |
| Vision-and-Language Navigation | REVERIE (val unseen) | SPL25.2 | 129 | |
| Vision-and-Language Navigation | R4R unseen (val) | Success Rate (SR)20.5 | 52 | |
| Room-to-Room Navigation | R2R 72 scenes | NE5.86 | 5 |