Equivariant Diffusion Policy
About
Recent work has shown diffusion models are an effective approach to learning the multimodal distributions arising from demonstration data in behavior cloning. However, a drawback of this approach is the need to learn a denoising function, which is significantly more complex than learning an explicit policy. In this work, we propose Equivariant Diffusion Policy, a novel diffusion policy learning method that leverages domain symmetries to obtain better sample efficiency and generalization in the denoising function. We theoretically analyze the $\mathrm{SO}(2)$ symmetry of full 6-DoF control and characterize when a diffusion model is $\mathrm{SO}(2)$-equivariant. We furthermore evaluate the method empirically on a set of 12 simulation tasks in MimicGen, and show that it obtains a success rate that is, on average, 21.9% higher than the baseline Diffusion Policy. We also evaluate the method on a real-world system to show that effective policies can be learned with relatively few training samples, whereas the baseline Diffusion Policy cannot.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Coffee Making/Handling | Robomimic MimicGen Coffee (D2) | Success Rate65 | 25 | |
| Coffee Preparation | Robomimic/MimicGen Coffee Prep. (D1) | Success Rate80 | 20 | |
| Mug Cleanup | Robomimic MimicGen Mug Cleanup (D1) | Success Rate54 | 20 | |
| Robot Manipulation | MimicGen Square D2 | Success Rate39 | 15 | |
| Robot Manipulation | MimicGen Nut Assembly D0 | Success Rate72 | 15 | |
| Robotic Manipulation | MimicGen SE(2) | Stack (D1) Success Rate99 | 11 | |
| Robot Manipulation | MimicGen Hammer Cleanup D1 | Success Rate70 | 10 | |
| Robot Manipulation | MimicGen Stack D1 | Success Rate98 | 10 | |
| Robot Manipulation | MimicGen Stack Three D1 | Success Rate76 | 10 | |
| Robot Manipulation Policy Inference | MimicGen | Coffee Success Rate (D2)2.44 | 8 |