Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Context Learning for Multi-Agent Discussion

About

Multi-Agent Discussion (MAD) has garnered increasing attention very recently, where multiple LLM instances collaboratively solve problems via structured discussion. However, we find that current MAD methods easily suffer from discussion inconsistency, LLMs fail to reach a coherent solution, due to the misalignment between their individual contexts.In this paper, we introduce a multi-LLM context learning method (M2CL) that learns a context generator for each agent, capable of dynamically generating context instructions per discussion round via automatic information organization and refinement. Specifically, inspired by our theoretical insights on the context instruction, M2CL train the generators to control context coherence and output discrepancies via a carefully crafted self-adaptive mechanism.It enables LLMs to avoid premature convergence on majority noise and progressively reach the correct consensus. We evaluate M2CL on challenging tasks, including academic reasoning, embodied tasks, and mobile control. The results show that the performance of M2CL significantly surpasses existing methods by 20%--50%, while enjoying favorable transferability and computational efficiency.

Xingyuan Hua, Sheng Yue, Xinyi Li, Yizhe Zhao, Jinrui Zhang, Ju Ren• 2026

Related benchmarks

TaskDatasetResultRank
Multi-task Language UnderstandingMMLU
Accuracy99.7
842
Language UnderstandingMMLU
Accuracy96.6
756
Mathematical ReasoningMATH
Accuracy74.7
643
General AI Assistant TasksGAIA
Accuracy84.2
266
Question AnsweringGPQA
Accuracy78.7
258
Code GenerationCode
Accuracy98.7
242
Automated PlanningPDDL
Accuracy83.9
233
Science ReasoningGPQA
Accuracy95.1
218
Mathematical Problem SolvingMATH
Accuracy79.7
166
Scientific ReasoningSciworld
Accuracy95.9
164
Showing 10 of 23 rows

Other info

GitHub

Follow for update