Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Fed-SE: Federated Self-Evolution for Privacy-Constrained Multi-Environment LLM Agents

About

LLM agents are widely deployed in complex interactive tasks, yet privacy constraints often preclude centralized optimization and co-evolution across dynamic environments. Despite the demonstrated success of Federated Learning (FL) on static datasets, its effectiveness in open-ended, self-evolving agent systems remains largely unexplored. In such settings, the direct application of standard FL is particularly challenging, as heterogeneous tasks and sparse, trajectory-level reward signals give rise to severe gradient instability, which undermines the global optimization process. To bridge this gap, we propose Fed-SE, a Federated Self-Evolution framework for LLM agents that establishes a local evolution-global aggregation paradigm. Locally, agents employ parameter-efficient fine-tuning on filtered, high-return trajectories to achieve stable gradient updates. Globally, Fed-SE aggregates updates within a low-rank subspace, reducing communication cost across clients. Experiments across five heterogeneous environments demonstrate that Fed-SE improves average task success rates by 10\% over the state-of-the-art FedIT, validating its effectiveness in cross-environment knowledge transfer under privacy constraints.

Xiang Chen, Yuling Shi, Qizhen Lan, Yuchao Qiu, Min Wang, Xiaodong Gu, Yanfu Yan• 2025

Related benchmarks

TaskDatasetResultRank
Item craftingTextCraft (test)
Success Rate71
32
Game SolvingWordle (test)
Success Rate32
25
LLM Agent NavigationBabyAI (test)
Success Rate93.3
25
Maze NavigationMaze (test)
Success Rate0.8
25
Multi-Task Agent GeneralizationFive Agent Environments Average (test)
Average Success Rate73.2
25
Web-based Agent InteractionWebShop (test)
Success Rate73
25
Showing 6 of 6 rows

Other info

Follow for update