Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Learning Human-to-Robot Handovers from Point Clouds

About

We propose the first framework to learn control policies for vision-based human-to-robot handovers, a critical task for human-robot interaction. While research in Embodied AI has made significant progress in training robot agents in simulated environments, interacting with humans remains challenging due to the difficulties of simulating humans. Fortunately, recent research has developed realistic simulated environments for human-to-robot handovers. Leveraging this result, we introduce a method that is trained with a human-in-the-loop via a two-stage teacher-student framework that uses motion and grasp planning, reinforcement learning, and self-supervision. We show significant performance gains over baselines on a simulation benchmark, sim-to-sim transfer and sim-to-real transfer.

Sammy Christen, Wei Yang, Claudia P\'erez-D'Arpino, Otmar Hilliges, Dieter Fox, Yu-Wei Chao• 2023

Related benchmarks

TaskDatasetResultRank
Human-to-Robot HandoverHandoverSim Simultaneous setting s0 (test)
Success Rate (%)68.75
15
Human-to-Robot HandoverHandoverSim s0 (Sequential) v1 (test)
Success Rate75.23
13
Robot HandoverHandoverSim Sequential (s0)
Success Rate (S)75.23
13
Human-to-Robot HandoverGenH2R-Sim t0 v1 (test)
Success Score (S)33.71
12
Robot HandoverGenH2R-Sim (t0)
Success Rate33.71
12
Human-to-Robot HandoverGenH2R-Sim t1 v1 (test)
Success Rate52.4
12
Robot HandoverGenH2R-Sim (t1)
Success Rate (%)52.4
12
Robot HandoverHandoverSim Simultaneous (s0)
Success Rate (S)68.75
12
HandoverHandoverSim S2: Unseen Handedness
Success Rate72.96
8
Human-robot handoverHandoverSim (S1: Unseen Subjects)
Success Rate (%)75
8
Showing 10 of 13 rows

Other info

Code

Follow for update