Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

The Option-Critic Architecture

About

Temporal abstraction is key to scaling up learning and planning in reinforcement learning. While planning with temporally extended actions is well understood, creating such abstractions autonomously from data has remained challenging. We tackle this problem in the framework of options [Sutton, Precup & Singh, 1999; Precup, 2000]. We derive policy gradient theorems for options and propose a new option-critic architecture capable of learning both the internal policies and the termination conditions of options, in tandem with the policy over options, and without the need to provide any additional rewards or subgoals. Experimental results in both discrete and continuous environments showcase the flexibility and efficiency of the framework.

Pierre-Luc Bacon, Jean Harb, Doina Precup• 2016

Related benchmarks

TaskDatasetResultRank
SlidingFetchSlide
Success Rate39
18
pushingFetchPush
Success Rate11
18
Pick-&-PlaceFetchPickAndPlace
Success Rate7
18
ReachingFetchReach
Success Rate0.00e+0
12
Reinforcement LearningGrid World Npick=3 Dense (test)
Max Average Return2
2
Reinforcement LearningGrid World Npick=5, Sparse (test)
Maximum Average Return0.00e+0
2
Showing 6 of 6 rows

Other info

Follow for update