Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Routing-Free Mixture-of-Experts

About

Standard Mixture-of-Experts (MoE) models rely on centralized routing mechanisms that introduce rigid inductive biases. We propose Routing-Free MoE which eliminates any hard-coded centralized designs including external routers, Softmax, Top-K and load balancing, instead encapsulating all activation functionalities within individual experts and directly optimized through continuous gradient flow, enabling each expert to determine its activation entirely on its own. We introduce a unified adaptive load-balancing framework to simultaneously optimize both expert-balancing and token-balancing objectives through a configurable interpolation, allowing flexible and customizable resource allocation. Extensive experiments show that Routing-Free MoE can consistently outperform baselines with better scalability and robustness. We analyze its behavior in detail and offer insights that may facilitate future MoE design ad optimization.

Yilun Liu, Jinru Han, Sikuan Yan, Volker Tresp, Yunpu Ma• 2026

Related benchmarks

TaskDatasetResultRank
Commonsense ReasoningHellaSwag--
1891
Commonsense ReasoningWinoGrande
Accuracy50.59
1085
Question AnsweringARC Challenge--
906
Question AnsweringOpenBookQA
Normalized Accuracy26.6
102
Language ModelingOpenWebText
Perplexity19.97
91
Question AnsweringARC Easy
Normalized Accuracy37.46
18
Commonsense ReasoningPIQA
Normalized Accuracy58.92
13
Natural Language UnderstandingGLUE
QQP Accuracy39.93
8
Showing 8 of 8 rows

Other info

Follow for update