Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Autoregressive Knowledge Distillation through Imitation Learning

About

The performance of autoregressive models on natural language generation tasks has dramatically improved due to the adoption of deep, self-attentive architectures. However, these gains have come at the cost of hindering inference speed, making state-of-the-art models cumbersome to deploy in real-world, time-sensitive settings. We develop a compression technique for autoregressive models that is driven by an imitation learning perspective on knowledge distillation. The algorithm is designed to address the exposure bias problem. On prototypical language generation tasks such as translation and summarization, our method consistently outperforms other distillation algorithms, such as sequence-level knowledge distillation. Student models trained with our method attain 1.4 to 4.8 BLEU/ROUGE points higher than those trained from scratch, while increasing inference speed by up to 14 times in comparison to the teacher model.

Alexander Lin, Jeremy Wohlwend, Howard Chen, Tao Lei• 2020

Related benchmarks

TaskDatasetResultRank
Instruction FollowingUnNI
Rouge-L28.7
160
Instruction FollowingS-NI
Rouge-L33.1
119
Instruction FollowingDollyEval
Rouge-L25.3
114
Instruction FollowingVicuna
Rouge-L16
83
Commonsense ReasoningStrategyQA (test)
Accuracy61.7
81
Instruction FollowingSelfInst
Rouge-L18.4
73
Instruction FollowingVicunaEval
Rouge-L19.1
72
Abstractive dialogue summarizationSamSum (test)
ROUGE-L51.2
53
Mathematical ReasoningGSM-Plus (test)
Accuracy21.3
50
Text SummarizationDialogueSUM (test)
ROUGE-L35.1
49
Showing 10 of 23 rows

Other info

Code

Follow for update