Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RMA: Rapid Motor Adaptation for Legged Robots

About

Successful real-world deployment of legged robots would require them to adapt in real-time to unseen scenarios like changing terrains, changing payloads, wear and tear. This paper presents Rapid Motor Adaptation (RMA) algorithm to solve this problem of real-time online adaptation in quadruped robots. RMA consists of two components: a base policy and an adaptation module. The combination of these components enables the robot to adapt to novel situations in fractions of a second. RMA is trained completely in simulation without using any domain knowledge like reference trajectories or predefined foot trajectory generators and is deployed on the A1 robot without any fine-tuning. We train RMA on a varied terrain generator using bioenergetics-inspired rewards and deploy it on a variety of difficult terrains including rocky, slippery, deformable surfaces in environments with grass, long vegetation, concrete, pebbles, stairs, sand, etc. RMA shows state-of-the-art performance across diverse real-world as well as simulation experiments. Video results at https://ashish-kmr.github.io/rma-legged-robots/

Ashish Kumar, Zipeng Fu, Deepak Pathak, Jitendra Malik• 2021

Related benchmarks

TaskDatasetResultRank
LocomotionHopper IID (test)
Mean Episode Reward1.86e+3
24
Locomotion ControlAnt sigma 0.3 (test)
Episode Reward1.72e+3
24
Locomotion ControlAnt sigma 0.5 (test)
Episode Reward974
24
LocomotionWalker IID (test)
Mean Episode Reward1.78e+3
24
Locomotion ControlWalker sigma 0.1 (test)
Episode Reward1.78e+3
24
Locomotion ControlAnt sigma 0.1 (test)
Episode Reward2.15e+3
24
LocomotionAnt IID (test)
Mean Episode Reward2.15e+3
24
LocomotionHalf Cheetah IID (test)
Mean Episode Reward2.20e+3
24
Locomotion ControlHalf Cheetah sigma 0.3 (test)
Episode Reward1.47e+3
24
Locomotion ControlHopper sigma 0.3 (test)
Episode Reward1.30e+3
24
Showing 10 of 30 rows

Other info

Follow for update