SPOC: Imitating Shortest Paths in Simulation Enables Effective Navigation and Manipulation in the Real World
About
Reinforcement learning (RL) with dense rewards and imitation learning (IL) with human-generated trajectories are the most widely used approaches for training modern embodied agents. RL requires extensive reward shaping and auxiliary losses and is often too slow and ineffective for long-horizon tasks. While IL with human supervision is effective, collecting human trajectories at scale is extremely expensive. In this work, we show that imitating shortest-path planners in simulation produces agents that, given a language instruction, can proficiently navigate, explore, and manipulate objects in both simulation and in the real world using only RGB sensors (no depth map or GPS coordinates). This surprising result is enabled by our end-to-end, transformer-based, SPOC architecture, powerful visual encoders paired with extensive image augmentation, and the dramatic scale and diversity of our training data: millions of frames of shortest-path-expert trajectories collected inside approximately 200,000 procedurally generated houses containing 40,000 unique 3D assets. Our models, data, training code, and newly proposed 10-task benchmarking suite CHORES are available in https://spoc-robot.github.io.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Fetch (FETCH) | Real-world | Success Rate44.4 | 5 | |
| Object Navigation | OBJNAV | Success Rate57.5 | 5 | |
| Object Navigation | OBJNAVROOM | Success Rate50.3 | 5 | |
| Object Navigation | OBJNAVRELATTR | Success Rate54.6 | 5 | |
| Object Navigation | OBJNAVAFFORD | Success Rate62.4 | 5 | |
| Object Navigation (OBJNAV) | Real-world | Success Rate83.3 | 5 | |
| Pickup (PICKUP) | Real-world | Success Rate46.7 | 5 | |
| Room Visit (ROOMVISIT) | Real-world | Success Rate50 | 5 |