Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RvS: What is Essential for Offline RL via Supervised Learning?

About

Recent work has shown that supervised learning alone, without temporal difference (TD) learning, can be remarkably effective for offline RL. When does this hold true, and which algorithmic components are necessary? Through extensive experiments, we boil supervised learning for offline RL down to its essential elements. In every environment suite we consider, simply maximizing likelihood with a two-layer feedforward MLP is competitive with state-of-the-art results of substantially more complex methods based on TD learning or sequence modeling with Transformers. Carefully choosing model capacity (e.g., via regularization or architecture) and choosing which information to condition on (e.g., goals or rewards) are critical for performance. These insights serve as a field guide for practitioners doing Reinforcement Learning via Supervised Learning (which we coin "RvS learning"). They also probe the limits of existing RvS methods, which are comparatively weak on random data, and suggest a number of open problems.

Scott Emmons, Benjamin Eysenbach, Ilya Kostrikov, Sergey Levine• 2021

Related benchmarks

TaskDatasetResultRank
Offline Reinforcement LearningKitchen Partial
Normalized Score71.7
62
hopper locomotionD4RL hopper medium-replay
Normalized Score73.5
56
walker2d locomotionD4RL walker2d medium-replay
Normalized Score60.6
53
LocomotionD4RL walker2d-medium-expert
Normalized Score106
47
LocomotionD4RL Halfcheetah medium
Normalized Score42.6
44
LocomotionD4RL Walker2d medium
Normalized Score0.717
44
Offline Reinforcement LearningD4RL antmaze-umaze (diverse)
Normalized Score66.2
40
hopper locomotionD4RL Hopper medium
Normalized Score60.2
38
hopper locomotionD4RL hopper-medium-expert
Normalized Score101.7
38
LocomotionD4RL halfcheetah-medium-expert
Normalized Score92.2
37
Showing 10 of 54 rows

Other info

Follow for update