RMA: Rapid Motor Adaptation for Legged Robots
About
Successful real-world deployment of legged robots would require them to adapt in real-time to unseen scenarios like changing terrains, changing payloads, wear and tear. This paper presents Rapid Motor Adaptation (RMA) algorithm to solve this problem of real-time online adaptation in quadruped robots. RMA consists of two components: a base policy and an adaptation module. The combination of these components enables the robot to adapt to novel situations in fractions of a second. RMA is trained completely in simulation without using any domain knowledge like reference trajectories or predefined foot trajectory generators and is deployed on the A1 robot without any fine-tuning. We train RMA on a varied terrain generator using bioenergetics-inspired rewards and deploy it on a variety of difficult terrains including rocky, slippery, deformable surfaces in environments with grass, long vegetation, concrete, pebbles, stairs, sand, etc. RMA shows state-of-the-art performance across diverse real-world as well as simulation experiments. Video results at https://ashish-kmr.github.io/rma-legged-robots/
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Locomotion | Hopper IID (test) | Mean Episode Reward1.86e+3 | 24 | |
| Locomotion Control | Ant sigma 0.3 (test) | Episode Reward1.72e+3 | 24 | |
| Locomotion Control | Ant sigma 0.5 (test) | Episode Reward974 | 24 | |
| Locomotion | Walker IID (test) | Mean Episode Reward1.78e+3 | 24 | |
| Locomotion Control | Walker sigma 0.1 (test) | Episode Reward1.78e+3 | 24 | |
| Locomotion Control | Ant sigma 0.1 (test) | Episode Reward2.15e+3 | 24 | |
| Locomotion | Ant IID (test) | Mean Episode Reward2.15e+3 | 24 | |
| Locomotion | Half Cheetah IID (test) | Mean Episode Reward2.20e+3 | 24 | |
| Locomotion Control | Half Cheetah sigma 0.3 (test) | Episode Reward1.47e+3 | 24 | |
| Locomotion Control | Hopper sigma 0.3 (test) | Episode Reward1.30e+3 | 24 |