Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

EasyMimic: A Low-Cost Framework for Robot Imitation Learning from Human Videos

About

Robot imitation learning is often hindered by the high cost of collecting large-scale, real-world data. This challenge is especially significant for low-cost robots designed for home use, as they must be both user-friendly and affordable. To address this, we propose the EasyMimic framework, a low-cost and replicable solution that enables robots to quickly learn manipulation policies from human video demonstrations captured with standard RGB cameras. Our method first extracts 3D hand trajectories from the videos. An action alignment module then maps these trajectories to the gripper control space of a low-cost robot. To bridge the human-to-robot domain gap, we introduce a simple and user-friendly hand visual augmentation strategy. We then use a co-training method, fine-tuning a model on both the processed human data and a small amount of robot data, enabling rapid adaptation to new tasks. Experiments on the low-cost LeRobot platform demonstrate that EasyMimic achieves high performance across various manipulation tasks. It significantly reduces the reliance on expensive robot data collection, offering a practical path for bringing intelligent robots into homes. Project website: https://zt375356.github.io/EasyMimic-Project/.

Tao Zhang, Song Xia, Ye Wang, Qin Jin• 2026

Related benchmarks

TaskDatasetResultRank
Language-conditioned manipulationEasyMimic Dataset 1.0 (test)
Average Score90
4
PickEasyMimic Dataset 1.0 (test)
Average Score1
4
StackEasyMimic Dataset 1.0 (test)
Average Score0.7
4
PullEasyMimic Dataset 1.0 (test)
Avg Score0.9
4
Pick-&-PlacePick and Place Unseen Objects: green duck, pink cube
Success Rate (Green Duck)80
2
Showing 5 of 5 rows

Other info

Follow for update