Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Policy Regularization with Dataset Constraint for Offline Reinforcement Learning

About

We consider the problem of learning the best possible policy from a fixed dataset, known as offline Reinforcement Learning (RL). A common taxonomy of existing offline RL works is policy regularization, which typically constrains the learned policy by distribution or support of the behavior policy. However, distribution and support constraints are overly conservative since they both force the policy to choose similar actions as the behavior policy when considering particular states. It will limit the learned policy's performance, especially when the behavior policy is sub-optimal. In this paper, we find that regularizing the policy towards the nearest state-action pair can be more effective and thus propose Policy Regularization with Dataset Constraint (PRDC). When updating the policy in a given state, PRDC searches the entire dataset for the nearest state-action sample and then restricts the policy with the action of this sample. Unlike previous works, PRDC can guide the policy with proper behaviors from the dataset, allowing it to choose actions that do not appear in the dataset along with the given state. It is a softer constraint but still keeps enough conservatism from out-of-distribution actions. Empirical evidence and theoretical analysis show that PRDC can alleviate offline RL's fundamentally challenging value overestimation issue with a bounded performance gap. Moreover, on a set of locomotion and navigation tasks, PRDC achieves state-of-the-art performance compared with existing methods. Code is available at https://github.com/LAMDA-RL/PRDC

Yuhang Ran, Yi-Chen Li, Fuxiang Zhang, Zongzhang Zhang, Yang Yu• 2023

Related benchmarks

TaskDatasetResultRank
hopper locomotionD4RL hopper medium-replay
Normalized Score100.1
56
walker2d locomotionD4RL walker2d medium-replay
Normalized Score92
53
LocomotionD4RL walker2d-medium-expert
Normalized Score111.2
47
LocomotionD4RL Halfcheetah medium
Normalized Score63.5
44
LocomotionD4RL Walker2d medium
Normalized Score85.2
44
hopper locomotionD4RL Hopper medium
Normalized Score100.3
38
hopper locomotionD4RL hopper-medium-expert
Normalized Score109.2
38
LocomotionD4RL halfcheetah-medium-expert
Normalized Score94.5
37
LocomotionD4RL HalfCheetah Medium-Replay
Normalized Score0.55
33
LocomotionD4RL halfcheetah-random-medium
Normalized Score56.5
5
Showing 10 of 18 rows

Other info

Follow for update