Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

DistilQwen2.5: Industrial Practices of Training Distilled Open Lightweight Language Models

About

Enhancing computational efficiency and reducing deployment costs for large language models (LLMs) have become critical challenges in various resource-constrained scenarios. In this work, we present DistilQwen2.5, a family of distilled, lightweight LLMs derived from the public Qwen2.5 models. These distilled models exhibit enhanced instruction-following capabilities compared to the original models based on a series of distillation techniques that incorporate knowledge from much larger LLMs. In our industrial practice, we first leverage powerful proprietary LLMs with varying capacities as multi-agent teachers to select, rewrite, and refine instruction-response pairs that are more suitable for student LLMs to learn. After standard fine-tuning, we further leverage a computationally efficient model fusion approach that enables student models to progressively integrate fine-grained hidden knowledge from their teachers. Experimental evaluations demonstrate that the distilled models possess significantly stronger capabilities than their original checkpoints. Additionally, we present use cases to illustrate the applications of our framework in real-world scenarios. To facilitate practical use, we have released all the DistilQwen2.5 models to the open-source community.

Chengyu Wang, Junbing Yan, Yuanhao Yue, Jun Huang• 2025

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningMATH
Accuracy13.48
643
Code GenerationMBPP
Pass@142.91
175
Mathematical ReasoningGSM8K
Math Score58.44
171
Code GenerationHumanEval
Pass@134.71
108
Instruction FollowingDollyEval
Score40.62
106
Agentic Reasoning∞Bench
Score59.22
100
Code GenerationLiveCodeBench
Pass@123.84
86
Instruction FollowingVicunaEval
VicunaEval Score38.99
80
Code GenerationLiveCodeBench
Average Score26.76
68
Code GenerationMBPP
Score46.88
38
Showing 10 of 14 rows

Other info

Follow for update