Dynamic Mix Precision Routing for Efficient Multi-step LLM Interaction
About
Large language models (LLM) achieve strong performance in long-horizon decision-making tasks through multi-step interaction and reasoning at test time. While practitioners commonly believe a higher task success rate necessitates the use of a larger and stronger LLM model, multi-step interaction with a large LLM incurs prohibitive inference cost. To address this problem, we explore the use of low-precision quantized LLM in the long-horizon decision-making process. Based on the observation of diverse sensitivities among interaction steps, we propose a dynamic mix-precision routing framework that adaptively selects between high-precision and low-precision LLMs at each decision step. The router is trained via a two-stage pipeline, consisting of KL-divergence-based supervised learning that identifies precision-sensitive steps, followed by Group-Relative Policy Optimization (GRPO) to further improve task success rates. Experiments on ALFWorld demonstrate that our approach achieves a great improvement on accuracy-cost trade-off over single-precision baselines and heuristic routing methods.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-step LLM Interaction | Qwen3-8B Inference Performance (test) | High-Precision Ratio26.7 | 7 | |
| Multi-step LLM Interaction | Qwen3-1.7B Inference Performance (test) | High-Precision Ratio20.5 | 7 | |
| Multi-step LLM Interaction | Qwen3-4B Inference Performance (test) | High-Precision Ratio8.6 | 7 | |
| Multi-step LLM Interaction | DeepSeek-R1-Distill-Llama-8B Inference Performance (test) | High-Precision Ratio9.8 | 7 |