Towards Fast Multilingual LLM Inference: Speculative Decoding and Specialized Drafters
About
Large language models (LLMs) have revolutionized natural language processing and broadened their applicability across diverse commercial applications. However, the deployment of these models is constrained by high inference time in multilingual settings. To mitigate this challenge, this paper explores a training recipe of an assistant model in speculative decoding, which is leveraged to draft and-then its future tokens are verified by the target LLM. We show that language-specific draft models, optimized through a targeted pretrain-and-finetune strategy, substantially brings a speedup in inference time compared to the previous methods. We validate these models across various languages in inference time, out-of-domain speedup, and GPT-4o evaluation.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Code Generation | MT-Bench (test) | Speedup Ratio2.437 | 26 | |
| Machine Translation | WMT German-English 16 (test) | Speedup ratio2.076 | 26 | |
| Question Answering | Natural Questions (test) | Speedup Ratio1.96 | 26 | |
| Summarization | CNN/Daily Mail (test) | Speedup Ratio2.133 | 26 | |
| Mathematical Reasoning | GSM8K (test) | Relative Speedup2.454 | 17 | |
| Machine Translation | JA-EN | Speedup Ratio1.757 | 8 | |
| Machine Translation | Ru-En | Speedup Ratio1.817 | 8 | |
| Machine Translation | DE-EN | Speedup Ratio2.36 | 8 | |
| Machine Translation | FR-EN | Speedup Ratio2.135 | 8 | |
| Machine Translation | Zh-En | Speedup Ratio1.516 | 8 |