When to Trust Your Model: Model-Based Policy Optimization
About
Designing effective model-based reinforcement learning algorithms is difficult because the ease of data generation must be weighed against the bias of model-generated data. In this paper, we study the role of model usage in policy optimization both theoretically and empirically. We first formulate and analyze a model-based reinforcement learning algorithm with a guarantee of monotonic improvement at each step. In practice, this analysis is overly pessimistic and suggests that real off-policy data is always preferable to model-generated on-policy data, but we show that an empirical estimate of model generalization can be incorporated into such analysis to justify model usage. Motivated by this analysis, we then demonstrate that a simple procedure of using short model-generated rollouts branched from real data has the benefits of more complicated model-based algorithms without the usual pitfalls. In particular, this approach surpasses the sample efficiency of prior model-based methods, matches the asymptotic performance of the best model-free algorithms, and scales to horizons that cause other model-based methods to fail entirely.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Goal Reaching | RoboKitchen (test) | Success Rate22.2 | 16 | |
| Goal Reaching | fetch_push (test) | Success Rate0.719 | 10 | |
| Goal Reaching | maze_large (test) | Success Rate6 | 10 | |
| Goal Reaching | pinpad (test) | Average Success Rate6.8 | 10 | |
| UAV Trajectory Planning | Agricultural application scenario random map (test) | Reward1.64e+3 | 9 | |
| UAV Trajectory Planning | Urban application scenario random map (test) | Reward134.1 | 9 | |
| Continuous Control | InvertedDouble v2 | Average Return9.36e+3 | 7 | |
| Off-dynamics Reinforcement Learning | Reacher 0.5 density v1 (test) | Reward-11.7 | 7 | |
| Reinforcement Learning | Ant 0.5 gravity (test) | Average Return981 | 7 | |
| Continuous Control | HalfCheetah v3 | Average Return6.78e+3 | 7 |