An Imitation from Observation Approach to Transfer Learning with Dynamics Mismatch
About
We examine the problem of transferring a policy learned in a source environment to a target environment with different dynamics, particularly in the case where it is critical to reduce the amount of interaction with the target environment during learning. This problem is particularly important in sim-to-real transfer because simulators inevitably model real-world dynamics imperfectly. In this paper, we show that one existing solution to this transfer problem - grounded action transformation - is closely related to the problem of imitation from observation (IfO): learning behaviors that mimic the observations of behavior demonstrations. After establishing this relationship, we hypothesize that recent state-of-the-art approaches from the IfO literature can be effectively repurposed for grounded transfer learning.To validate our hypothesis we derive a new algorithm - generative adversarial reinforced action transformation (GARAT) - based on adversarial imitation from observation techniques. We run experiments in several domains with mismatched dynamics, and find that agents trained with GARAT achieve higher returns in the target environment compared to existing black-box transfer methods
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Off-dynamics Reinforcement Learning | Ant MuJoCo | Average Reward3.38e+3 | 7 | |
| Reinforcement Learning | HalfCheetah 1.5 gravity MuJoCo | Reward3.83e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Walker2d MuJoCo | Average Reward3.30e+3 | 7 | |
| Reinforcement Learning | Walker2d 0.5 gravity (test) | Average Return823 | 7 | |
| Reinforcement Learning | Ant 1.5 gravity MuJoCo | Reward1.96e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Ant 0.5 density v1 (test) | Reward2.15e+3 | 7 | |
| Reinforcement Learning | HalfCheetah 0.5 gravity (test) | Average Return3.44e+3 | 7 | |
| Reinforcement Learning | Reacher 1.5 gravity MuJoCo | Reward-16.7 | 7 | |
| Reinforcement Learning | MuJoCo Reacher 1.5 density v1 (test) | Reward-13.3 | 7 | |
| Off-dynamics Reinforcement Learning | HalfCheetah broken source environment MuJoCo | Average Reward5.88e+3 | 7 |