Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Actor-Attention-Critic for Multi-Agent Reinforcement Learning

About

Reinforcement learning in multi-agent scenarios is important for real-world applications but presents challenges beyond those seen in single-agent settings. We present an actor-critic algorithm that trains decentralized policies in multi-agent settings, using centrally computed critics that share an attention mechanism which selects relevant information for each agent at every timestep. This attention mechanism enables more effective and scalable learning in complex multi-agent environments, when compared to recent approaches. Our approach is applicable not only to cooperative settings with shared rewards, but also individualized reward settings, including adversarial settings, as well as settings that do not provide global states, and it makes no assumptions about the action spaces of the agents. As such, it is flexible enough to be applied to most multi-agent learning problems.

Shariq Iqbal, Fei Sha• 2018

Related benchmarks

TaskDatasetResultRank
Multi-Agent Reinforcement LearningTREA rdist
Mean Episodic Reward1.73e+4
42
Multi-Agent Reinforcement LearningCN rdete
Mean Episodic Reward-227
21
Multi-Agent Reinforcement LearningREF rdete
Mean Episodic Reward-69
21
Multi-Agent Reinforcement LearningCN rdist
Mean Episodic Reward-270
21
Multi-Agent Reinforcement LearningREF rdist
Mean Episodic Reward-72
21
Multi-Agent Reinforcement LearningTREA rdete
Mean Episodic Reward-480
21
Multi-Agent Reinforcement LearningCN rac-dist
Mean Episodic Reward-85
21
Multi-Agent Reinforcement LearningREF rac-dist
Mean Episodic Reward215
21
Multi-Agent Reinforcement LearningCN-q rac-dist (test)
Mean Episodic Reward (q=3)132
12
Multi-Agent Reinforcement LearningREF-q rac-dist (test)
Mean Episodic Reward (q=2)11
12
Showing 10 of 11 rows

Other info

Follow for update