Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MiniLLM: On-Policy Distillation of Large Language Models

About

Knowledge Distillation (KD) is a promising technique for reducing the high computational demand of large language models (LLMs). However, previous KD methods are primarily applied to white-box classification models or training small models to imitate black-box model APIs like ChatGPT. How to effectively distill the knowledge of white-box LLMs into small models is still under-explored, which becomes more important with the prosperity of open-source LLMs. In this work, we propose a KD approach that distills LLMs into smaller language models. We first replace the forward Kullback-Leibler divergence (KLD) objective in the standard KD approaches with reverse KLD, which is more suitable for KD on generative language models, to prevent the student model from overestimating the low-probability regions of the teacher distribution. Then, we derive an effective on-policy optimization approach to learn this objective. The student models are named MiniLLM. Extensive experiments in the instruction-following setting show that MiniLLM generates more precise responses with higher overall quality, lower exposure bias, better calibration, and higher long-text generation performance than the baselines. Our method is scalable for different model families with 120M to 13B parameters. Our code, data, and model checkpoints can be found in https://github.com/microsoft/LMOps/tree/main/minillm.

Yuxian Gu, Li Dong, Furu Wei, Minlie Huang• 2023

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Accuracy57.7
1362
Mathematical ReasoningMATH
Accuracy22.6
882
ReasoningBBH
Accuracy36
672
Instruction FollowingIFEval
IFEval Accuracy65.8
625
Logical reasoningBBH
Accuracy36
201
Instruction FollowingUnNI
Rouge-L40.2
160
Code GenerationMBPP
Accuracy42.2
159
Science Question AnsweringSciQ
Normalized Accuracy84.1
137
Instruction FollowingS-NI
Rouge-L35.5
119
Instruction FollowingDollyEval
Rouge-L29
114
Showing 10 of 25 rows

Other info

Code

Follow for update