Any-point Trajectory Modeling for Policy Learning
About
Learning from demonstration is a powerful method for teaching robots new skills, and having more demonstration data often improves policy learning. However, the high cost of collecting demonstration data is a significant bottleneck. Videos, as a rich data source, contain knowledge of behaviors, physics, and semantics, but extracting control-specific information from them is challenging due to the lack of action labels. In this work, we introduce a novel framework, Any-point Trajectory Modeling (ATM), that utilizes video demonstrations by pre-training a trajectory model to predict future trajectories of arbitrary points within a video frame. Once trained, these trajectories provide detailed control guidance, enabling the learning of robust visuomotor policies with minimal action-labeled data. Across over 130 language-conditioned tasks we evaluated in both simulation and the real world, ATM outperforms strong video pre-training baselines by 80% on average. Furthermore, we show effective transfer learning of manipulation skills from human videos and videos from a different robot morphology. Visualizations and code are available at: \url{https://xingyu-lin.github.io/atm}.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robot Manipulation | LIBERO | Goal Achievement79.6 | 700 | |
| Robot Manipulation | LIBERO (test) | Average Success Rate65.7 | 184 | |
| Open microwave | Simulation | Success Rate99.4 | 18 | |
| Block Stack | Simulation | Success Rate91.9 | 18 | |
| Glass | Simulation | Success Rate58.6 | 18 | |
| open drawer | Real-World (test) | Success Rate30 | 11 | |
| Robotic Manipulation | MetaWorld | Door Open Success Rate75.3 | 10 | |
| Motion forecasting | Panthera High motion (test) | Variance (Velocity)10 | 9 | |
| Motion forecasting | Panthera Combined (test) | Var (V)2.42 | 9 | |
| Text-conditioned trajectory prediction | LIBERO-90 | Side MSE47.82 | 8 |