Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Self-Paced Deep Reinforcement Learning

About

Curriculum reinforcement learning (CRL) improves the learning speed and stability of an agent by exposing it to a tailored series of tasks throughout learning. Despite empirical successes, an open question in CRL is how to automatically generate a curriculum for a given reinforcement learning (RL) agent, avoiding manual design. In this paper, we propose an answer by interpreting the curriculum generation as an inference problem, where distributions over tasks are progressively learned to approach the target task. This approach leads to an automatic curriculum generation, whose pace is controlled by the agent, with solid theoretical motivation and easily integrated with deep RL algorithms. In the conducted experiments, the curricula generated with the proposed algorithm significantly improve learning performance across several environments and deep RL algorithms, matching or outperforming state-of-the-art existing CRL algorithms.

Pascal Klink, Carlo D'Eramo, Jan Peters, Joni Pajarinen• 2020

Related benchmarks

TaskDatasetResultRank
Point Mass navigationPoint Mass environment setup 2
Average Collected Reward22.57
3
Ball CatchingBall Catching environment
Average Collected Reward (Mean)-51.11
3
Lunar Lander ControlLunar Lander
Average Collected Reward (Mean)252.6
3
Point Mass navigationPoint Mass environment (Setup 1)
Average Collected Reward22.16
3
Showing 4 of 4 rows

Other info

Follow for update