Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MASER: Multi-Agent Reinforcement Learning with Subgoals Generated from Experience Replay Buffer

About

In this paper, we consider cooperative multi-agent reinforcement learning (MARL) with sparse reward. To tackle this problem, we propose a novel method named MASER: MARL with subgoals generated from experience replay buffer. Under the widely-used assumption of centralized training with decentralized execution and consistent Q-value decomposition for MARL, MASER automatically generates proper subgoals for multiple agents from the experience replay buffer by considering both individual Q-value and total Q-value. Then, MASER designs individual intrinsic reward for each agent based on actionable representation relevant to Q-learning so that the agents reach their subgoals while maximizing the joint action value. Numerical results show that MASER significantly outperforms StarCraft II micromanagement benchmark compared to other state-of-the-art MARL algorithms.

Jeewon Jeon, Woojun Kim, Whiyoung Jung, Youngchul Sung• 2022

Related benchmarks

TaskDatasetResultRank
Multi-Agent Reinforcement LearningMAMuJoCo HalfCheetah 6x1 (test)
Average Episodic Return-3.14
8
Multi-Agent Reinforcement LearningMAMuJoCo Hopper 3x1 (test)
Average Episodic Return5.86
8
Multi-Agent Reinforcement LearningMAMuJoCo Ant 8x1 (test)
Average Episodic Return13.65
8
Multi-Agent Reinforcement LearningMAMuJoCo Walker2d 6x1 (test)
Average Episodic Return7.4
8
Showing 4 of 4 rows

Other info

Follow for update