Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Offline RL Without Off-Policy Evaluation

About

Most prior approaches to offline reinforcement learning (RL) have taken an iterative actor-critic approach involving off-policy evaluation. In this paper we show that simply doing one step of constrained/regularized policy improvement using an on-policy Q estimate of the behavior policy performs surprisingly well. This one-step algorithm beats the previously reported results of iterative algorithms on a large portion of the D4RL benchmark. The one-step baseline achieves this strong performance while being notably simpler and more robust to hyperparameters than previously proposed iterative algorithms. We argue that the relatively poor performance of iterative approaches is a result of the high variance inherent in doing off-policy evaluation and magnified by the repeated optimization of policies against those estimates. In addition, we hypothesize that the strong performance of the one-step algorithm is due to a combination of favorable structure in the environment and behavior policy.

David Brandfonbrener, William F. Whitney, Rajesh Ranganath, Joan Bruna• 2021

Related benchmarks

TaskDatasetResultRank
Offline Reinforcement LearningD4RL halfcheetah-medium-expert
Normalized Score93.5
117
Offline Reinforcement LearningD4RL hopper-medium-expert
Normalized Score102.1
115
Offline Reinforcement LearningD4RL walker2d-medium-expert
Normalized Score112.9
86
Offline Reinforcement LearningD4RL walker2d-random
Normalized Score6.1
77
Offline Reinforcement LearningD4RL halfcheetah-random
Normalized Score6.9
70
Offline Reinforcement LearningD4RL Walker2d Medium v2
Normalized Return81.8
67
Offline Reinforcement LearningD4RL hopper-random
Normalized Score7.8
62
Offline Reinforcement LearningD4RL halfcheetah v2 (medium-replay)
Normalized Score38.1
58
hopper locomotionD4RL hopper medium-replay
Normalized Score97.5
56
Offline Reinforcement LearningD4RL Hopper-medium-replay v2
Normalized Return97.5
54
Showing 10 of 63 rows

Other info

Code

Follow for update