Bridging Training and Execution via Dynamic Directed Graph-Based Communication in Cooperative Multi-Agent Systems
About
Multi-agent systems must learn to communicate and understand interactions between agents to achieve cooperative goals in partially observed tasks. However, existing approaches lack a dynamic directed communication mechanism and rely on global states, thus diminishing the role of communication in centralized training. Thus, we propose the Transformer-based graph coarsening network (TGCNet), a novel multi-agent reinforcement learning (MARL) algorithm. TGCNet learns the topological structure of a dynamic directed graph to represent the communication policy and integrates graph coarsening networks to approximate the representation of global state during training. It also utilizes the Transformer decoder for feature extraction during execution. Experiments on multiple cooperative MARL benchmarks demonstrate state-of-the-art performance compared to popular MARL algorithms. Further ablation studies validate the effectiveness of our dynamic directed graph communication mechanism and graph coarsening networks.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Cooperative Navigation | Cooperative Navigation easy | Mean Episode Reward2.08 | 14 | |
| Cooperative Navigation | multi-agent particle environment medium | Average Return-2.21 | 7 | |
| Cooperative Navigation | Cooperative Navigation hard | Mean Episode Reward-2.35 | 7 | |
| Multi-agent cooperation | SMAC 1o_2r_vs_4r (easy) | Win Rate63.28 | 7 | |
| Multi-agent cooperation | SMAC 1o_2r_vs_4r medium | Win Rate56.25 | 7 | |
| Multi-agent cooperation | SMAC 1o_2r_vs_4r hard | Win Rate50.78 | 7 | |
| Multi-agent cooperation | SMAC 1o_2r_vs_4r super_hard | Win Rate49.22 | 7 | |
| Multi-agent cooperation | SMAC 1o_10b_vs_1r (easy) | Win Rate66.41 | 7 | |
| Multi-agent cooperation | SMAC 1o_10b_vs_1r medium | Win Rate51.92 | 7 | |
| Multi-agent cooperation | SMAC 1o_10b_vs_1r hard | Win Rate51.53 | 7 |