Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Constrained Ensemble Exploration for Unsupervised Skill Discovery

About

Unsupervised Reinforcement Learning (RL) provides a promising paradigm for learning useful behaviors via reward-free per-training. Existing methods for unsupervised RL mainly conduct empowerment-driven skill discovery or entropy-based exploration. However, empowerment often leads to static skills, and pure exploration only maximizes the state coverage rather than learning useful behaviors. In this paper, we propose a novel unsupervised RL framework via an ensemble of skills, where each skill performs partition exploration based on the state prototypes. Thus, each skill can explore the clustered area locally, and the ensemble skills maximize the overall state coverage. We adopt state-distribution constraints for the skill occupancy and the desired cluster for learning distinguishable skills. Theoretical analysis is provided for the state entropy and the resulting skill distributions. Based on extensive experiments on several challenging tasks, we find our method learns well-explored ensemble skills and achieves superior performance in various downstream tasks compared to previous methods.

Chenjia Bai, Rushuai Yang, Qiaosheng Zhang, Kang Xu, Yi Chen, Ting Xiao, Xuelong Li• 2024

Related benchmarks

TaskDatasetResultRank
State ExplorationMaze2D Square-b
State Coverage Ratio66
22
State ExplorationMaze2D Square-a
State Coverage Ratio71
11
State ExplorationMaze2D Square-c
State Coverage Ratio60
11
State ExplorationMaze2D Square-d
State Coverage Ratio0.57
11
State ExplorationMaze2D Corridor2
State Coverage Ratio82
11
State ExplorationMaze2D Square-tree
State Coverage Ratio40
11
Robotic ManipulationMetaWorld
Success Rate: Pick-out-of-hole0.00e+0
7
Showing 7 of 7 rows

Other info

Follow for update