Dueling Network Architectures for Deep Reinforcement Learning
About
In recent years there have been many successes of using deep representations in reinforcement learning. Still, many of these applications use conventional architectures, such as convolutional networks, LSTMs, or auto-encoders. In this paper, we present a new neural network architecture for model-free reinforcement learning. Our dueling network represents two separate estimators: one for the state value function and one for the state-dependent action advantage function. The main benefit of this factoring is to generalize learning across actions without imposing any change to the underlying reinforcement learning algorithm. Our results show that this architecture leads to better policy evaluation in the presence of many similar-valued actions. Moreover, the dueling architecture enables our RL agent to outperform the state-of-the-art on the Atari 2600 domain.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reinforcement Learning | Atari 2600 MONTEZUMA'S REVENGE | Score22 | 45 | |
| Atari Game Playing | Pitfall! | Score0.00e+0 | 25 | |
| Reinforcement Learning | Atari 2600 Montezuma's Revenge ALE (test) | Score0.00e+0 | 24 | |
| Reinforcement Learning | Atari 57 | Atlantis3.83e+5 | 21 | |
| Reinforcement Learning | Atari 2600 57 games | Median Human-Normalized Score151 | 20 | |
| Reinforcement Learning | Atari 2600 Private Eye ALE (test) | Score206 | 19 | |
| Reinforcement Learning | Atari 2600 Gravitar ALE (test) | Score238 | 19 | |
| Reinforcement Learning | Atari-57 (test) | Median Human Norm Return132 | 15 | |
| Reinforcement Learning | Atari 2600 57 games (test) | Median Human-Normalized Score172 | 15 | |
| Reinforcement Learning | Atari 2600 Freeway ALE (test) | Score33 | 14 |