Dual Latent Memory for Visual Multi-agent System
About
While Visual Multi-Agent Systems (VMAS) promise to enhance comprehensive abilities through inter-agent collaboration, empirical evidence reveals a counter-intuitive "scaling wall": increasing agent turns often degrades performance while exponentially inflating token costs. We attribute this failure to the information bottleneck inherent in text-centric communication, where converting perceptual and thinking trajectories into discrete natural language inevitably induces semantic loss. To this end, we propose L$^{2}$-VMAS, a novel model-agnostic framework that enables inter-agent collaboration with dual latent memories. Furthermore, we decouple the perception and thinking while dynamically synthesizing dual latent memories. Additionally, we introduce an entropy-driven proactive triggering that replaces passive information transmission with efficient, on-demand memory access. Extensive experiments among backbones, sizes, and multi-agent structures demonstrate that our method effectively breaks the "scaling wall" with superb scalability, improving average accuracy by 2.7-5.4% while reducing token usage by 21.3-44.8%. Codes: https://github.com/YU-deep/L2-VMAS.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multimodal Understanding | MMBench | Accuracy88.8 | 367 | |
| Video Understanding | MVBench | Accuracy73.1 | 247 | |
| Multimodal Understanding | MMStar | Accuracy81.4 | 197 | |
| Visual Question Answering | RealworldQA | Accuracy80.2 | 98 | |
| Visual Perception | BLINK | Accuracy72.7 | 71 | |
| Long Video Understanding | LVBench | Accuracy61.5 | 63 | |
| Multi-image Reasoning | MuirBench | Accuracy77.2 | 48 |