Model-Based Offline Planning
About
Offline learning is a key part of making reinforcement learning (RL) useable in real systems. Offline RL looks at scenarios where there is data from a system's operation, but no direct access to the system when learning a policy. Recent work on training RL policies from offline data has shown results both with model-free policies learned directly from the data, or with planning on top of learnt models of the data. Model-free policies tend to be more performant, but are more opaque, harder to command externally, and less easy to integrate into larger systems. We propose an offline learner that generates a model that can be used to control the system directly through planning. This allows us to have easily controllable policies directly from data, without ever interacting with the system. We show the performance of our algorithm, Model-Based Offline Planning (MBOP) on a series of robotics-inspired tasks, and demonstrate its ability leverage planning to respect environmental constraints. We are able to find near-optimal polices for certain simulated systems from as little as 50 seconds of real-time system interaction, and create zero-shot goal-conditioned policies on a series of environments. An accompanying video can be found here: https://youtu.be/nxGGHdZOFts
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | D4RL hopper-medium-expert | Normalized Score55.1 | 115 | |
| Offline Reinforcement Learning | D4RL walker2d-medium-expert | Normalized Score70.2 | 86 | |
| Offline Reinforcement Learning | D4RL Walker2d Medium v2 | Normalized Return41 | 67 | |
| Offline Reinforcement Learning | D4RL HalfCheetah Medium v2 | Average Normalized Return44.6 | 43 | |
| Offline Reinforcement Learning | D4RL Hopper Medium v2 | Normalized Return48.8 | 43 | |
| Offline Reinforcement Learning | D4RL Hopper medium | Reward48.8 | 35 | |
| Offline Reinforcement Learning | D4RL Gym walker2d medium-expert | Normalized Average Return70.2 | 31 | |
| Offline Reinforcement Learning | D4RL hopper medium-replay | Reward12.4 | 30 | |
| Offline Reinforcement Learning | D4RL HalfCheetah Med-Replay v2 | Avg Normalized Return42.3 | 29 | |
| Offline Reinforcement Learning | D4RL Halfcheetah medium | Reward44.6 | 28 |