Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

SAM Decoding: Speculative Decoding via Suffix Automaton

About

Speculative decoding (SD) has been demonstrated as an effective technique for lossless LLM inference acceleration. Retrieval-based SD methods, one kind of model-free method, have yielded promising speedup, but they often rely on incomplete retrieval resources, inefficient retrieval methods, and are constrained to certain domains. This paper presents a novel retrieval-based speculative decoding method that adapts suffix automaton (SAM) for efficient and accurate draft generation by utilizing common text corpus and dynamic text sequence. Unlike existing $n$-gram matching methods, SAM-Decoding finds the exact longest suffix match, achieving an average time complexity of O(1) per generation step of SAM update and suffix retrieval. It can also integrate with existing methods, adaptively selecting a draft generation strategy based on match length to generalize to broader domains. Extensive experiments on Spec-Bench show that our method is $18\%+$ faster than other retrieval-based SD methods. Additionally, when combined with advanced EAGLE-2, it provides an additional speedup of $3.28\%$ -- $11.13\%$ across various-sized LLM backbones. Our code is available at our \href{https://github.com/hyx1999/SAM-Decoding}{repository}.

Yuxuan Hu, Ke Wang, Xiaokang Zhang, Fanjin Zhang, Cuiping Li, Hong Chen, Jing Zhang• 2024

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Tokens/s91.5
61
Inference EfficiencyHumanEval
Speedup Factor3.35
54
Speculative DecodingSpec-Bench
MT Score4.62
48
Inference AccelerationSpec-Bench
MAT Score4.62
39
Inference EfficiencyHAGRID
#MAT4.75
9
Information RetrievalHAGRID
#MAT4.41
6
Question AnsweringHAGRID
Match Count (#MAT)3.93
6
Retrieval-Augmented GenerationHAGRID
#MAT3.82
6
Showing 8 of 8 rows

Other info

Follow for update