Benchmarking Batch Deep Reinforcement Learning Algorithms
About
Widely-used deep reinforcement learning algorithms have been shown to fail in the batch setting--learning from a fixed data set without interaction with the environment. Following this result, there have been several papers showing reasonable performances under a variety of environments and batch settings. In this paper, we benchmark the performance of recent off-policy and batch reinforcement learning algorithms under unified settings on the Atari domain, with data generated by a single partially-trained behavioral policy. We find that under these conditions, many of these algorithms underperform DQN trained online with the same amount of data, as well as the partially-trained behavioral policy. To introduce a strong baseline, we adapt the Batch-Constrained Q-learning algorithm to a discrete-action setting, and show it outperforms all existing algorithms at this task.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Sudoku Solving | Sudoku 2x2 | Final Reward1.3 | 14 | |
| Constrained Reinforcement Learning | GRID | Episodic Reward276.3 | 8 | |
| Constrained Reinforcement Learning | Bottleneck | Episodic Reward298.3 | 8 | |
| Constrained Reinforcement Learning | AntReach | Episodic Reward54.2 | 8 | |
| Constrained Reinforcement Learning | AntCircle | Episodic Reward134.4 | 8 | |
| Constrained Reinforcement Learning | Humanoid | Episodic Reward1.43e+3 | 8 | |
| Constrained Reinforcement Learning | PointCircle | Episodic Reward57.2 | 8 | |
| Constrained Reinforcement Learning | PointReach | Episodic Reward46.1 | 8 | |
| Stabilization | Safe-Control-Gym Quadrotor Stab Observation Uncertainty | Average Return164 | 7 | |
| Tracking | Safe-Control-Gym Quadrotor Track Observation Uncertainty | Average Return218 | 7 |