LoRA-Drop: Temporal LoRA Decoding for Efficient LLM Inference
About
Autoregressive large language models (LLMs) are bottlenecked by sequential decoding, where each new token typically requires executing all transformer layers. Existing dynamic-depth and layer-skipping methods reduce this cost, but often rely on auxiliary routing mechanisms or incur accuracy degradation when bypassed layers are left uncompensated. We present \textbf{LoRA-Drop}, a plug-and-play inference framework that accelerates decoding by applying a \emph{temporal compute schedule} to a fixed subset of intermediate layers: on most decoding steps, selected layers reuse the previous-token hidden state and apply a low-rank LoRA correction, while periodic \emph{refresh} steps execute the full model to prevent drift. LoRA-Drop requires no routing network, is compatible with standard KV caching, and can reduce KV-cache footprint by skipping KV updates in droppable layers during LoRA steps and refreshing periodically. Across \textbf{LLaMA2-7B}, \textbf{LLaMA3-8B}, \textbf{Qwen2.5-7B}, and \textbf{Qwen2.5-14B}, LoRA-Drop achieves up to \textbf{2.6$\times$ faster decoding} and \textbf{45--55\% KV-cache reduction} while staying within \textbf{0.5 percentage points (pp)} of baseline accuracy. Evaluations on reasoning (GSM8K, MATH, BBH), code generation (HumanEval, MBPP), and long-context/multilingual benchmarks (LongBench, XNLI, XCOPA) identify a consistent \emph{safe zone} of scheduling configurations that preserves quality while delivering substantial efficiency gains, providing a simple path toward adaptive-capacity inference in LLMs. Codes are available at https://github.com/hosseinbv/LoRA-Drop.git.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Commonsense Reasoning | HellaSwag | Accuracy77 | 1460 | |
| Code Generation | HumanEval | Pass@140.7 | 850 | |
| Multi-task Language Understanding | MMLU | -- | 842 | |
| Commonsense Reasoning | WinoGrande | Accuracy73.6 | 776 | |
| Commonsense Reasoning | PIQA | Accuracy81.1 | 647 | |
| Mathematical Reasoning | MATH | Accuracy16.5 | 535 | |
| Reasoning | BBH | Accuracy39 | 507 | |
| Question Answering | ARC Easy | Normalized Acc80.2 | 385 | |
| Language Modeling | LAMBADA | Accuracy74.6 | 183 | |
| Natural Language Inference | XNLI | Accuracy69 | 111 |