Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Less is More: Task-aware Layer-wise Distillation for Language Model Compression

About

Layer-wise distillation is a powerful tool to compress large models (i.e. teacher models) into small ones (i.e., student models). The student distills knowledge from the teacher by mimicking the hidden representations of the teacher at every intermediate layer. However, layer-wise distillation is difficult. Since the student has a smaller model capacity than the teacher, it is often under-fitted. Furthermore, the hidden representations of the teacher contain redundant information that the student does not necessarily need for the target task's learning. To address these challenges, we propose a novel Task-aware layEr-wise Distillation (TED). TED designs task-aware filters to align the hidden representations of the student and the teacher at each layer. The filters select the knowledge that is useful for the target task from the hidden representations. As such, TED reduces the knowledge gap between the two models and helps the student to fit better on the target task. We evaluate TED in two scenarios: continual pre-training and fine-tuning. TED demonstrates significant and consistent improvements over existing distillation methods in both scenarios. Code is available at https://github.com/cliang1453/task-aware-distillation.

Chen Liang, Simiao Zuo, Qingru Zhang, Pengcheng He, Weizhu Chen, Tuo Zhao• 2022

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K (test)
Accuracy49.28
751
Code GenerationHumanEval (test)--
444
Multitask Language UnderstandingMMLU (test)
Accuracy45.21
303
Code GenerationMBPP (test)--
276
Science Question AnsweringARC Challenge
Accuracy55.47
234
Multiple-choice Question AnsweringMMLU
Accuracy45.47
148
Science Question AnsweringARC-E
Accuracy74.94
138
Commonsense ReasoningWinoGrande (val)
Accuracy57.38
87
Commonsense ReasoningCommonsenseQA (val)
Accuracy65.27
52
Science Question AnsweringARC Easy
Accuracy28.61
26
Showing 10 of 20 rows

Other info

Follow for update