Deep Imitation Learning for Complex Manipulation Tasks from Virtual Reality Teleoperation
About
Imitation learning is a powerful paradigm for robot skill acquisition. However, obtaining demonstrations suitable for learning a policy that maps from raw pixels to actions can be challenging. In this paper we describe how consumer-grade Virtual Reality headsets and hand tracking hardware can be used to naturally teleoperate robots to perform complex tasks. We also describe how imitation learning can learn deep neural network policies (mapping from pixels to actions) that can acquire the demonstrated skills. Our experiments showcase the effectiveness of our approach for learning visuomotor skills.
Tianhao Zhang, Zoe McCarthy, Owen Jow, Dennis Lee, Xi Chen, Ken Goldberg, Pieter Abbeel• 2017
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Bimanual Insertion | Bimanual Insertion sim | Grasp Success10 | 10 | |
| Cube Transfer | Cube Transfer sim | Touched Count34 | 10 | |
| Slide Ziploc | Slide Ziploc real | Grasp Success0.00e+0 | 5 | |
| Slot Battery | Slot Battery real | Grasp Success0.00e+0 | 5 |
Showing 4 of 4 rows