CityWalker: Learning Embodied Urban Navigation from Web-Scale Videos
About
Navigating dynamic urban environments presents significant challenges for embodied agents, requiring advanced spatial reasoning and adherence to common-sense norms. Despite progress, existing visual navigation methods struggle in map-free or off-street settings, limiting the deployment of autonomous agents like last-mile delivery robots. To overcome these obstacles, we propose a scalable, data-driven approach for human-like urban navigation by training agents on thousands of hours of in-the-wild city walking and driving videos sourced from the web. We introduce a simple and scalable data processing pipeline that extracts action supervision from these videos, enabling large-scale imitation learning without costly annotations. Our model learns sophisticated navigation policies to handle diverse challenges and critical scenarios. Experimental results show that training on large-scale, diverse datasets significantly enhances navigation performance, surpassing current methods. This work shows the potential of using abundant online video data to develop robust navigation policies for embodied agents in dynamic urban settings. Project homepage is at https://ai4ce.github.io/CityWalker/.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Trajectory Prediction | StereoWalker Benchmark Offline | Mean L2 (m)0.75 | 6 | |
| Semantic Navigation | Simulation Environment Small Gazebo | SR (%)100 | 5 | |
| Semantic Navigation | Simulation Environment Large Gazebo | SR (%)80 | 5 | |
| Point-Goal navigation | CityWalker (test) | MAOE (Mean)15.2 | 5 | |
| Point-Goal navigation | SocNav Benchmark | SR47.8 | 5 | |
| Semantic Navigation | Simulation Environment Medium Gazebo | Success Rate0.85 | 5 | |
| POI-Goal Navigation | BridgeNav Dataset (test) | SR (0.1m)13.79 | 4 | |
| Vision-Language Navigation | BridgeNav | Success Rate (0.1m)13.79 | 4 |