Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DHRL: A Graph-Based Approach for Long-Horizon and Sparse Hierarchical Reinforcement Learning

About

Hierarchical Reinforcement Learning (HRL) has made notable progress in complex control tasks by leveraging temporal abstraction. However, previous HRL algorithms often suffer from serious data inefficiency as environments get large. The extended components, $i.e.$, goal space and length of episodes, impose a burden on either one or both high-level and low-level policies since both levels share the total horizon of the episode. In this paper, we present a method of Decoupling Horizons Using a Graph in Hierarchical Reinforcement Learning (DHRL) which can alleviate this problem by decoupling the horizons of high-level and low-level policies and bridging the gap between the length of both horizons using a graph. DHRL provides a freely stretchable high-level action interval, which facilitates longer temporal abstraction and faster training in complex tasks. Our method outperforms state-of-the-art HRL algorithms in typical HRL environments. Moreover, DHRL achieves long and complex locomotion and manipulation tasks.

Seungjae Lee, Jigang Kim, Inkyu Jang, H. Jin Kim• 2022

Related benchmarks

TaskDatasetResultRank
NavigationPointMaze
Success Rate9.98e+3
21
NavigationAntMaze Small
Success Rate9.51e+3
16
NavigationAntMaze
Success Rate9.11e+3
16
NavigationBottleneck
Success Rate38.7
16
NavigationComplex
Success Rate4.01e+3
16
ReachingReacher 3D
Success Rate95.1
10
Obstacle AvoidanceUR3Obstacle
Success Rate69.8
8
Showing 7 of 7 rows

Other info

Follow for update