Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

RODE: Learning Roles to Decompose Multi-Agent Tasks

About

Role-based learning holds the promise of achieving scalable multi-agent learning by decomposing complex tasks using roles. However, it is largely unclear how to efficiently discover such a set of roles. To solve this problem, we propose to first decompose joint action spaces into restricted role action spaces by clustering actions according to their effects on the environment and other agents. Learning a role selector based on action effects makes role discovery much easier because it forms a bi-level learning hierarchy -- the role selector searches in a smaller role space and at a lower temporal resolution, while role policies learn in significantly reduced primitive action-observation spaces. We further integrate information about action effects into the role policies to boost learning efficiency and policy generalization. By virtue of these advances, our method (1) outperforms the current state-of-the-art MARL algorithms on 10 of the 14 scenarios that comprise the challenging StarCraft II micromanagement benchmark and (2) achieves rapid transfer to new environments with three times the number of agents. Demonstrative videos are available at https://sites.google.com/view/rode-marl .

Tonghan Wang, Tarun Gupta, Anuj Mahajan, Bei Peng, Shimon Whiteson, Chongjie Zhang• 2020

Related benchmarks

TaskDatasetResultRank
Multi-Agent Reinforcement LearningStarCraft Multi-Agent Challenge (SMAC)
1c3s5z Win Rate100
13
Multi-Agent Reinforcement LearningSMAC 3s5z_vs_3s6z v1 (test)
Win Rate96.8
9
Multi-Agent Reinforcement LearningSMAC 2s_vs_1sc v1 (test)
Win Rate100
9
Multi-Agent Reinforcement LearningSMAC corridor v1 (test)
Win Rate65.6
9
Showing 4 of 4 rows

Other info

Follow for update