Model-Based Episodic Memory Induces Dynamic Hybrid Controls
About
Episodic control enables sample efficiency in reinforcement learning by recalling past experiences from an episodic memory. We propose a new model-based episodic memory of trajectories addressing current limitations of episodic control. Our memory estimates trajectory values, guiding the agent towards good policies. Built upon the memory, we construct a complementary learning model via a dynamic hybrid control unifying model-based, episodic and habitual learning into a single architecture. Experiments demonstrate that our model allows significantly faster and better learning than other strong reinforcement learning agents across a variety of environments including stochastic and non-Markovian settings.
Hung Le, Thommen Karimpanal George, Majid Abdolshah, Truyen Tran, Svetha Venkatesh• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Reinforcement Learning | Atari 2600 57 games | Median Human-Normalized Score117.2 | 20 | |
| Reinforcement Learning | Atari 2600 25 games | Mean Human Normalized Score518.2 | 7 | |
| Reinforcement Learning | Atari 2600 (10M frames) | Alien Score1.99e+3 | 2 |
Showing 3 of 3 rows