Imitation Learning via Off-Policy Distribution Matching
About
When performing imitation learning from expert demonstrations, distribution matching is a popular approach, in which one alternates between estimating distribution ratios and then using these ratios as rewards in a standard reinforcement learning (RL) algorithm. Traditionally, estimation of the distribution ratio requires on-policy data, which has caused previous work to either be exorbitantly data-inefficient or alter the original objective in a manner that can drastically change its optimum. In this work, we show how the original distribution ratio estimation objective may be transformed in a principled manner to yield a completely off-policy objective. In addition to the data-efficiency that this provides, we are able to show that this objective also renders the use of a separate RL optimization unnecessary.Rather, an imitation policy may be learned directly from this objective without the use of explicit rewards. We call the resulting algorithm ValueDICE and evaluate it on a suite of popular imitation learning benchmarks, finding that it can achieve state-of-the-art sample efficiency and performance.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | D4RL hopper-expert v2 | Normalized Score65.6 | 56 | |
| Offline Reinforcement Learning | D4RL walker2d-expert v2 | Normalized Score28.2 | 56 | |
| Offline Reinforcement Learning | D4RL halfcheetah-expert v2 | Normalized Score9.8 | 56 | |
| Offline Imitation Learning | D4RL Ant v2 (expert) | Normalized Score90.5 | 20 | |
| Continuous Control | MuJoCo Ant | Average Reward4.51e+3 | 12 | |
| Continuous Control | MuJoCo HalfCheetah | Average Reward4.84e+3 | 12 | |
| Robotic Manipulation | Robomimic Lift | Success Rate47.6 | 12 | |
| Robotic Manipulation | Robomimic Can | Success Rate41.8 | 12 | |
| Robotic Manipulation | Robomimic Square | Success Rate8.3 | 12 | |
| Action-matching | MIMIC-III (test) | Accuracy79.4 | 9 |