Unsupervised Domain Adaptation with Dynamics-Aware Rewards in Reinforcement Learning
About
Unsupervised reinforcement learning aims to acquire skills without prior goal representations, where an agent automatically explores an open-ended environment to represent goals and learn the goal-conditioned policy. However, this procedure is often time-consuming, limiting the rollout in some potentially expensive target environments. The intuitive approach of training in another interaction-rich environment disrupts the reproducibility of trained skills in the target environment due to the dynamics shifts and thus inhibits direct transferring. Assuming free access to a source environment, we propose an unsupervised domain adaptation method to identify and acquire skills across dynamics. Particularly, we introduce a KL regularized objective to encourage emergence of skills, rewarding the agent for both discovering skills and aligning its behaviors respecting dynamics shifts. This suggests that both dynamics (source and target) shape the reward to facilitate the learning of adaptive skills. We also conduct empirical experiments to demonstrate that our method can effectively learn skills that can be smoothly deployed in target.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Off-dynamics Reinforcement Learning | Walker2d 0.5 density v1 (test) | Reward1.93e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Reacher 0.5 density v1 (test) | Reward-12.2 | 7 | |
| Reinforcement Learning | MuJoCo HalfCheetah 1.5 density v1 (test) | Reward5.06e+3 | 7 | |
| Off-dynamics Reinforcement Learning | HalfCheetah broken source environment MuJoCo | Average Reward6.40e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Ant MuJoCo | Average Reward3.24e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Reacher broken source environment MuJoCo | Average Reward-13.9 | 7 | |
| Off-dynamics Reinforcement Learning | Ant 0.5 density v1 (test) | Reward2.23e+3 | 7 | |
| Reinforcement Learning | HalfCheetah 0.5 gravity (test) | Average Return3.67e+3 | 7 | |
| Reinforcement Learning | Reacher 1.5 gravity MuJoCo | Reward-16.5 | 7 | |
| Reinforcement Learning | MuJoCo Reacher 1.5 density v1 (test) | Reward-11.3 | 7 |