Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FLEX-MoE: Federated Mixture-of-Experts with Load-balanced Expert Assignment

About

Mixture-of-Experts (MoE) models enable scalable neural networks through conditional computation. However, their deployment with federated learning (FL) faces two critical challenges: 1) resource-constrained edge devices cannot store full expert sets, and 2) non-IID data distributions cause severe expert load imbalance that degrades model performance. To this end, we propose \textbf{FLEX-MoE}, a novel federated MoE framework that jointly optimizes expert assignment and load balancing under limited client capacity. Specifically, our approach introduces client-expert fitness scores that quantify the expert suitability for local datasets through training feedback, and employs an optimization-based algorithm to maximize client-expert specialization while enforcing balanced expert utilization system-wide. Unlike existing greedy methods that focus solely on personalization while ignoring load imbalance, our FLEX-MoE is capable of addressing the expert utilization skew, which is particularly severe in FL settings with heterogeneous data. Our comprehensive experiments on three different datasets demonstrate the superior performance of the proposed FLEX-MoE, together with its ability to maintain balanced expert utilization across diverse resource-constrained scenarios.

Boyang Zhang, Xiaobing Chen, Songyang Zhang, Shuai Zhang, Xiangwei Zhou, Mingxuan Sun• 2025

Related benchmarks

TaskDatasetResultRank
Image ClassificationCIFAR10 0.1-Dirichlet (test)--
38
Image ClassificationEMNIST (Non-IID (4-Class partition))
Accuracy82.57
12
Image ClassificationCIFAR10 IID
Accuracy83
5
Image ClassificationEMNIST (IID)
Accuracy0.9833
5
Image ClassificationGTSRB (IID)
Accuracy (%)98
5
Image ClassificationCIFAR10 Non-IID (Dirichlet alpha = 0.8)
Accuracy0.6881
5
Image ClassificationEMNIST Non-IID (Dirichlet alpha = 0.8)
Accuracy86.91
5
Image ClassificationGTSRB Non-IID (Dirichlet alpha = 0.8)
Accuracy76.5
5
Image ClassificationCIFAR10 Non-IID (4-Class partition)
Accuracy65.51
5
Image ClassificationCIFAR10 Non-IID 2-Class partition (test)
Accuracy61
5
Showing 10 of 11 rows

Other info

Follow for update