Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

bert2BERT: Towards Reusable Pretrained Language Models

About

In recent years, researchers tend to pre-train ever-larger language models to explore the upper limit of deep models. However, large language model pre-training costs intensive computational resources and most of the models are trained from scratch without reusing the existing pre-trained models, which is wasteful. In this paper, we propose bert2BERT, which can effectively transfer the knowledge of an existing smaller pre-trained model (e.g., BERT_BASE) to a large model (e.g., BERT_LARGE) through parameter initialization and significantly improve the pre-training efficiency of the large model. Specifically, we extend the previous function-preserving on Transformer-based language model, and further improve it by proposing advanced knowledge for large model's initialization. In addition, a two-stage pre-training method is proposed to further accelerate the training process. We did extensive experiments on representative PLMs (e.g., BERT and GPT) and demonstrate that (1) our method can save a significant amount of training cost compared with baselines including learning from scratch, StackBERT and MSLT; (2) our method is generic and applicable to different types of pre-trained models. In particular, bert2BERT saves about 45% and 47% computational cost of pre-training BERT_BASE and GPT_BASE by reusing the models of almost their half sizes. The source code will be publicly available upon publication.

Cheng Chen, Yichun Yin, Lifeng Shang, Xin Jiang, Yujia Qin, Fengyu Wang, Zhi Wang, Xiao Chen, Zhiyuan Liu, Qun Liu• 2021

Related benchmarks

TaskDatasetResultRank
Image ClassificationStanford Cars
Accuracy82.9
635
Image ClassificationFood-101
Accuracy82
542
Natural Language UnderstandingGLUE
SST-292.89
531
Image ClassificationCIFAR-100
Accuracy74.9
435
ClassificationCars
Accuracy91.88
395
Image ClassificationCUB-200 2011
Accuracy67.1
356
Image ClassificationCIFAR100
Accuracy90.47
347
Image ClassificationOxford Flowers 102
Accuracy85.4
234
Image ClassificationCIFAR10
Accuracy98.99
137
Image ClassificationFlowers
Accuracy97.51
127
Showing 10 of 19 rows

Other info

Follow for update