Dejavu: Towards Experience Feedback Learning for Embodied Intelligence
About
Embodied agents face a fundamental limitation: once deployed in real-world environments, they cannot easily acquire new knowledge to improve task performance. In this paper, we propose Dejavu, a general post-deployment learning framework that augments a frozen Vision-Language-Action (VLA) policy with retrieved execution memories through an Experience Feedback Network (EFN). EFN identifies contextually relevant prior action experiences and conditions action prediction on the retrieved guidance. We train EFN with reinforcement learning and semantic similarity rewards, encouraging the predicted actions to align with past behaviors under the current observation. During deployment, EFN continually expands its memory with new trajectories, enabling the agent to exhibit ``learning from experience.'' Experiments across diverse embodied tasks show that EFN improves adaptability, robustness, and success rates over frozen baselines. Our Project Page is https://dejavu2025.github.io/.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robot Manipulation | LIBERO Object | Success Rate98.5 | 70 | |
| Robotic Manipulation | LIBERO Long | Success Rate94.6 | 44 | |
| Robot Manipulation | LIBERO Average | Success Rate97.2 | 26 | |
| Robot Manipulation | LIBERO Spatial | Success Rate98.2 | 21 | |
| Robot Manipulation | LIBERO Goal | Success Rate97.6 | 21 | |
| Robot Manipulation | AgiBot-G1 BottlePlace | Success Rate82 | 7 | |
| Robot Manipulation | AgiBot-G1 ShelfSort | Success Rate74.7 | 7 | |
| Robot Manipulation | AgiBot-G1 StockLift | Success Rate65.3 | 7 | |
| Robot Manipulation | AgiBot-G1 DrawerStore | Success Rate58.7 | 7 | |
| Robot Manipulation | AgiBot-G1 Average | Success Rate70.2 | 7 |