Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Stabilizing Off-Policy Q-Learning via Bootstrapping Error Reduction

About

Off-policy reinforcement learning aims to leverage experience collected from prior policies for sample-efficient learning. However, in practice, commonly used off-policy approximate dynamic programming methods based on Q-learning and actor-critic methods are highly sensitive to the data distribution, and can make only limited progress without collecting additional on-policy data. As a step towards more robust off-policy algorithms, we study the setting where the off-policy experience is fixed and there is no further interaction with the environment. We identify bootstrapping error as a key source of instability in current methods. Bootstrapping error is due to bootstrapping from actions that lie outside of the training data distribution, and it accumulates via the Bellman backup operator. We theoretically analyze bootstrapping error, and demonstrate how carefully constraining action selection in the backup can mitigate it. Based on our analysis, we propose a practical algorithm, bootstrapping error accumulation reduction (BEAR). We demonstrate that BEAR is able to learn robustly from different off-policy distributions, including random and suboptimal demonstrations, on a range of continuous control tasks.

Aviral Kumar, Justin Fu, George Tucker, Sergey Levine• 2019

Related benchmarks

TaskDatasetResultRank
Offline Reinforcement LearningD4RL halfcheetah-medium-expert
Normalized Score53.4
117
Offline Reinforcement LearningD4RL hopper-medium-expert
Normalized Score96.3
115
Offline Reinforcement LearningD4RL walker2d-medium-expert
Normalized Score40.1
86
Offline Reinforcement LearningD4RL walker2d-random
Normalized Score11.1
77
Offline Reinforcement LearningD4RL halfcheetah-random
Normalized Score25.1
70
Offline Reinforcement LearningD4RL Walker2d Medium v2
Normalized Return59.8
67
Offline Reinforcement LearningD4RL hopper-random
Normalized Score31.4
62
Offline Reinforcement LearningD4RL halfcheetah v2 (medium-replay)
Normalized Score48.6
58
Offline Reinforcement LearningD4RL walker2d-expert v2
Normalized Score110.1
56
Offline Reinforcement LearningD4RL halfcheetah-expert v2
Normalized Score92.7
56
Showing 10 of 135 rows
...

Other info

Follow for update