ReKep: Spatio-Temporal Reasoning of Relational Keypoint Constraints for Robotic Manipulation
About
Representing robotic manipulation tasks as constraints that associate the robot and the environment is a promising way to encode desired robot behaviors. However, it remains unclear how to formulate the constraints such that they are 1) versatile to diverse tasks, 2) free of manual labeling, and 3) optimizable by off-the-shelf solvers to produce robot actions in real-time. In this work, we introduce Relational Keypoint Constraints (ReKep), a visually-grounded representation for constraints in robotic manipulation. Specifically, ReKep is expressed as Python functions mapping a set of 3D keypoints in the environment to a numerical cost. We demonstrate that by representing a manipulation task as a sequence of Relational Keypoint Constraints, we can employ a hierarchical optimization procedure to solve for robot actions (represented by a sequence of end-effector poses in SE(3)) with a perception-action loop at a real-time frequency. Furthermore, in order to circumvent the need for manual specification of ReKep for each new task, we devise an automated procedure that leverages large vision models and vision-language models to produce ReKep from free-form language instructions and RGB-D observations. We present system implementations on a wheeled single-arm platform and a stationary dual-arm platform that can perform a large variety of manipulation tasks, featuring multi-stage, in-the-wild, bimanual, and reactive behaviors, all without task-specific data or environment models. Website at https://rekep-robot.github.io/.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Robotic Garment Grasping | Real-world multi-illumination garment dataset | Success Rate66.6667 | 20 | |
| Visuomotor Manipulation | Robot Manipulation Tasks 1.0 (OOD) | Average Success Rate (OOD)31.1 | 13 | |
| Visuomotor Manipulation | Robot Manipulation Tasks in-distribution 1.0 (ID) | Average Success Rate42.2 | 13 | |
| Low-level policy sampling | 5 typical action models (Pick, Place, Open, Close, Toggle) | Success Rate42 | 7 | |
| Robotic Manipulation | 8 Real-world Tasks 20 repetitions (test) | Place Food Success Rate80 | 6 | |
| Pour | Real Robot Experiments | Success Rate20 | 6 | |
| Garment Grasping | MIGG Luminance 0 – 40 1.0 (test) | Grasp Success Ratio (Glove)0.4 | 5 | |
| Robotic Grasping | RealData Lu low luminance 0-20 1.0 | Grasping Success Rate0.4 | 5 | |
| Robotic Grasping | RealData medium-high luminance 40-60 1.0 | Grasping Success Rate9 | 5 | |
| Robotic Grasping | RealData high luminance 60-80 1.0 | Grasping Success Rate0.6667 | 5 |