Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Turning Trash into Treasure: Accelerating Inference of Large Language Models with Token Recycling

About

Massive parameters of LLMs have made inference latency a fundamental bottleneck. Speculative decoding represents a lossless approach to accelerate inference through a guess-and-verify paradigm. Some methods rely on additional architectures to guess draft tokens, which need extra training before use. Alternatively, retrieval-based training-free techniques build libraries from pre-existing corpora or by n-gram generation. However, they face challenges like large storage requirements, time-consuming retrieval, and limited adaptability. Observing that candidate tokens generated during the decoding process are likely to reoccur in future sequences, we propose Token Recycling. It stores candidate tokens in an adjacency matrix and employs a breadth-first-search (BFS)-like algorithm to construct a draft tree, which is then validated through tree attention. New candidate tokens from the decoding process are then used to update the matrix. Token Recycling requires \textless2MB of additional storage and achieves approximately 2x speedup across all sizes of LLMs. It significantly outperforms existing train-free methods by 30\% and even a widely recognized training method by 25\%.

Xianzhen Luo, Yixuan Wang, Qingfu Zhu, Zhiming Zhang, Xuanyu Zhang, Qing Yang, Dongliang Xu• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Speed Up (x)1.98
177
Instruction FollowingAlpaca
Speedup (x)1.83
63
Code GenerationHumanEval
Tokens/s71.49
61
Inference EfficiencyHumanEval
Speedup Factor2.15
54
Speculative DecodingSpec-Bench
MT Score2.74
48
Multi-turn dialogueMT-Bench
Speedup1.72
47
Inference AccelerationSpec-Bench
MAT Score2.83
39
SummarizationCNN/DM
M Score2.88
35
Code GenerationHumanEval
Functional Score M2.83
29
LLM Inference AccelerationMBPP
Speedup2.15
24
Showing 10 of 21 rows

Other info

Follow for update