Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction
About
Off-policy reinforcement learning aims to leverage experience collected from prior policies for sample-efficient learning. However, in practice, commonly used off-policy approximate dynamic programming methods based on Q-learning and actor-critic methods are highly sensitive to the data distribution, and can make only limited progress without collecting additional on-policy data. As a step towards more robust off-policy algorithms, we study the setting where the off-policy experience is fixed and there is no further interaction with the environment. We identify bootstrapping error as a key source of instability in current methods. Bootstrapping error is due to bootstrapping from actions that lie outside of the training data distribution, and it accumulates via the Bellman backup operator. We theoretically analyze bootstrapping error, and demonstrate how carefully constraining action selection in the backup can mitigate it. Based on our analysis, we propose a practical algorithm, bootstrapping error accumulation reduction (BEAR). We demonstrate that BEAR is able to learn robustly from different off-policy distributions, including random and suboptimal demonstrations, on a range of continuous control tasks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Offline Reinforcement Learning | D4RL halfcheetah-medium-expert | Normalized Score53.4 | 117 | |
| Offline Reinforcement Learning | D4RL hopper-medium-expert | Normalized Score96.3 | 115 | |
| Offline Reinforcement Learning | D4RL walker2d-medium-expert | Normalized Score40.1 | 86 | |
| Offline Reinforcement Learning | D4RL walker2d-random | Normalized Score11.1 | 77 | |
| Offline Reinforcement Learning | D4RL halfcheetah-random | Normalized Score25.1 | 70 | |
| Offline Reinforcement Learning | D4RL Walker2d Medium v2 | Normalized Return59.8 | 67 | |
| Offline Reinforcement Learning | D4RL hopper-random | Normalized Score31.4 | 62 | |
| Offline Reinforcement Learning | D4RL halfcheetah v2 (medium-replay) | Normalized Score48.6 | 58 | |
| Offline Reinforcement Learning | D4RL walker2d-expert v2 | Normalized Score110.1 | 56 | |
| Offline Reinforcement Learning | D4RL halfcheetah-expert v2 | Normalized Score92.7 | 56 |