Cut the Crap: An Economical Communication Pipeline for LLM-based Multi-Agent Systems
About
Recent advancements in large language model (LLM)-powered agents have shown that collective intelligence can significantly outperform individual capabilities, largely attributed to the meticulously designed inter-agent communication topologies. Though impressive in performance, existing multi-agent pipelines inherently introduce substantial token overhead, as well as increased economic costs, which pose challenges for their large-scale deployments. In response to this challenge, we propose an economical, simple, and robust multi-agent communication framework, termed $\texttt{AgentPrune}$, which can seamlessly integrate into mainstream multi-agent systems and prunes redundant or even malicious communication messages. Technically, $\texttt{AgentPrune}$ is the first to identify and formally define the \textit{communication redundancy} issue present in current LLM-based multi-agent pipelines, and efficiently performs one-shot pruning on the spatial-temporal message-passing graph, yielding a token-economic and high-performing communication topology. Extensive experiments across six benchmarks demonstrate that $\texttt{AgentPrune}$ \textbf{(I)} achieves comparable results as state-of-the-art topologies at merely $\$5.6$ cost compared to their $\$43.7$, \textbf{(II)} integrates seamlessly into existing multi-agent frameworks with $28.1\%\sim72.8\%\downarrow$ token reduction, and \textbf{(III)} successfully defend against two types of agent-based adversarial attacks with $3.5\%\sim10.8\%\uparrow$ performance boost.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Code Generation | HumanEval | Pass@186.8 | 850 | |
| Language Understanding | MMLU | Accuracy85.07 | 756 | |
| Mathematical Reasoning | GSM8K (test) | Accuracy94.89 | 751 | |
| Code Generation | HumanEval (test) | Pass@196.6 | 444 | |
| Mathematical Reasoning | GSM8K | Accuracy88.73 | 351 | |
| Code Generation | MBPP (test) | Pass@192.3 | 276 | |
| Mathematical Reasoning | AQUA | Accuracy80.51 | 132 | |
| Science Question Answering | ARC-C | Accuracy86.73 | 127 | |
| General Reasoning | MMLU | MMLU Accuracy84.3 | 126 | |
| Mathematical Reasoning | MultiArith | Accuracy94.65 | 116 |