Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

RadixMLP -- Intra-batch Deduplication for Causal Transformers

About

Batch inference workloads for causal transformer models frequently process sequences that share common prefixes, such as system prompts, few-shot examples, or shared queries. Standard inference engines treat each sequence independently, redundantly recomputing identical MLP activations for every copy of the shared prefix. We introduce RadixMLP, a technique that exploits the position-wise nature of MLPs, LayerNorms, linear projections, and embeddings to eliminate this redundancy. RadixMLP dynamically maps batches to a prefix trie, gathering shared segments into a compressed representation for position-wise computation and scattering results back only at attention boundaries. RadixMLP is stateless and operates within a single forward pass. In end-to-end serving benchmarks on MS~MARCO v1.1 with Qwen3 models (0.6B to 8B parameters), RadixMLP achieves 1.44-1.59$\times$ speedups in realistic reranking workloads, with up to $5\times$ speedups on synthetic benchmarks with longer shared prefixes. Our code is available at https://github.com/michaelfeil/radix-mlp.

Michael Feil, Julius Lipp• 2026

Related benchmarks

TaskDatasetResultRank
InferenceMS MARCO b (test)
Latency P50 (s)0.55
9
InferenceMS MARCO c (test)
Latency p50 (s)0.57
9
Showing 2 of 2 rows

Other info

Follow for update