Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Distillation of Large Language Models via Concrete Score Matching

About

Large language models (LLMs) deliver remarkable performance but are costly to deploy, motivating knowledge distillation (KD) for efficient inference. Existing KD objectives typically match student and teacher probabilities via softmax, which blurs valuable logit information. While direct logit distillation (DLD) mitigates softmax smoothing, it fails to account for logit shift invariance, thereby restricting the solution space. We propose Concrete Score Distillation (CSD), a discrete score-matching objective that overcomes both softmax-induced smoothing and restrictions on the optimal solution set. We resolve the training instability and quadratic complexity of discrete score-matching in autoregressive LLMs, and the resulting CSD objective aligns relative logit differences across all vocabulary pairs between student and teacher with flexible weighting. We provide both mode-seeking and mode-covering instances within our framework and evaluate CSD on task-agnostic instruction-following and task-specific distillation using GPT-2-1.5B, OpenLLaMA-7B, and GEMMA-7B-IT. Experiments show that CSD consistently surpasses recent KD objectives, achieves favorable fidelity-diversity trade-offs, and yields complementary gains when combined with on-policy techniques, demonstrating its scalability and effectiveness for LLM distillation. Code: https://github.com/aailab-kaist/CSD.

Yeongmin Kim, Donghyeok Shin, Mina Kang, Byeonghu Na, Il-Chul Moon• 2025

Related benchmarks

TaskDatasetResultRank
Arithmetic ReasoningGSM8K
Accuracy25.78
173
Instruction FollowingUnNI
Rouge-L25.88
160
Instruction FollowingS-NI
Rouge-L24.6
119
Instruction FollowingDollyEval
Rouge-L24.94
114
ChatMT-Bench
MT-Bench Score7.85
58
Instruction FollowingSelf-Instruct
ROUGE-L12.06
48
Instruction FollowingSelf-Instruct (test)
ROUGE-L21.81
42
Instruction FollowingDolly Eval (test)
ROUGE-L29.63
42
ChatAlpacaEval
Win Rate90.05
39
Instruction FollowingVicuna Eval (test)
ROUGE-L20.37
36
Showing 10 of 15 rows

Other info

Follow for update