Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MoE-SpAc: Efficient MoE Inference Based on Speculative Activation Utility in Heterogeneous Edge Scenarios

About

Mixture-of-Experts (MoE) models enable scalable performance but face severe memory constraints on edge devices. Existing offloading strategies struggle with I/O bottlenecks due to the dynamic, low-information nature of autoregressive expert activation. In this paper, we propose to repurpose Speculative Decoding (SD) not merely as a compute accelerator, but as an informative lookahead sensor for memory management, supported by our theoretical and empirical analyses. Hence, we introduce MoE-SpAc, an MoE inference framework that integrates a Speculative Utility Estimator to track expert demand, a Heterogeneous Workload Balancer to dynamically partition computation via online integer optimization, and an Asynchronous Execution Engine to unify the prefetching and eviction in the same utility space. Extensive experiments on seven benchmarks demonstrate that MoE-SpAc achieves a 42% improvement in TPS over the SOTA SD-based baseline, and an average 4.04x speedup over all standard baselines. Code is available at https://github.com/lshAlgorithm/MoE-SpAc .

Shuhuai Li, Jianghao Lin, Dongdong Ge, Yinyu Ye• 2026

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K--
246
Instruction FollowingAlpaca--
111
Question AnsweringQA--
47
Text SummarizationCNN/DM
TPS50.17
13
Chat EvaluationMT-Bench
Throughput (TPS)24.1
10
Code GenerationHumanEval
TPS (Tokens/s)28.24
10
Language UnderstandingMMLU-Pro
TPS25.8
10
Code GenerationHumanEval
Throughput (tokens/s)47.26
6
Instruction FollowingMT-Bench
Throughput (TPS)47.1
3
Showing 9 of 9 rows

Other info

Follow for update