EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty
About
Autoregressive decoding makes the inference of Large Language Models (LLMs) time-consuming. In this paper, we reconsider speculative sampling and derive two key observations. Firstly, autoregression at the feature (second-to-top-layer) level is more straightforward than at the token level. Secondly, the inherent uncertainty in feature (second-to-top-layer) level autoregression constrains its performance. Based on these insights, we introduce EAGLE (Extrapolation Algorithm for Greater Language-model Efficiency), a simple yet highly efficient speculative sampling framework. By incorporating a token sequence advanced by one time step, EAGLE effectively resolves the uncertainty, enabling precise second-to-top-layer feature prediction with minimal overhead. We conducted comprehensive evaluations of EAGLE, including all models from the Vicuna and LLaMA2-Chat series, the MoE model Mixtral 8x7B Instruct, and tasks in dialogue, code generation, mathematical reasoning, and instruction following. For LLaMA2-Chat 70B, EAGLE achieved a latency speedup ratio of 2.7x-3.5x, doubled throughput, while maintaining the distribution of the generated text.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Mathematical Reasoning | GSM8K | Speed Up (x)1.5 | 177 | |
| Multi-turn dialogue | MT-Bench | Kendall's Tau3.93 | 54 | |
| Mathematical Reasoning | GSM8K | Tau ($ au$)4.01 | 54 | |
| Speculative Decoding | Spec-Bench | MT Score3.66 | 48 | |
| Code Generation | MBPP | Speed1.7 | 18 | |
| Mathematical Reasoning | MATH 500 | Speed1.4 | 18 | |
| Summarization | CNN/DM | Spd Score1.3 | 18 | |
| Code Generation | HumanEval | Speed (Spd)1.6 | 18 | |
| Autoregressive Image Generation | MS-COCO 2017 (val) | CLIP Score0.333 | 9 | |
| LLM Generation | SpecBench | Tokens/s106.9 | 3 |