Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SepLLM: Accelerate Large Language Models by Compressing One Segment into One Separator

About

Large Language Models (LLMs) have exhibited exceptional performance across a spectrum of natural language processing tasks. However, their substantial sizes pose considerable challenges, particularly in computational demands and inference speed, due to their quadratic complexity. In this work, we have identified a key pattern: certain seemingly meaningless separator tokens (i.e., punctuations) contribute disproportionately to attention scores compared to semantically meaningful tokens. This observation suggests that information of the segments between these separator tokens can be effectively condensed into the separator tokens themselves without significant information loss. Guided by this insight, we introduce SepLLM, a plug-and-play framework that accelerates inference by compressing these segments and eliminating redundant tokens. Additionally, we implement efficient kernels for training acceleration. Experimental results across training-free, training-from-scratch, and post-training settings demonstrate SepLLM's effectiveness. Notably, using the Llama-3-8B backbone, SepLLM achieves over 50% reduction in KV cache on the GSM8K-CoT benchmark while maintaining comparable performance. Furthermore, in streaming settings, SepLLM effectively processes sequences of up to 4 million tokens or more while maintaining consistent language modeling capabilities.

Guoxuan Chen, Han Shi, Jiawei Li, Yihang Gao, Xiaozhe Ren, Yimeng Chen, Xin Jiang, Zhenguo Li, Weiyang Liu, Chao Huang• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH500 (test)--
381
Mathematical ReasoningAIME 24
Accuracy33.3
113
Mathematical ReasoningAIME 2025 (test)
Pass@1 Rate31.3
47
Mathematical ReasoningGK EN 2023 (test)
Pass@172
16
Scientific Question AnsweringGPQA Diamond (test)
Pass@145.6
16
Mathematical ReasoningAverage (GSM8K, MATH-500, AMC23, AIME24, AIME25)
Accuracy61.1
14
Mathematical ReasoningAIME 25
Accuracy23.3
14
Mathematical ReasoningAMC 23
Accuracy80
12
Mathematical ReasoningGSM8K
Accuracy84.9
12
Mathematical ReasoningMATH 500
Accuracy84.2
12
Showing 10 of 10 rows

Other info

Follow for update