Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Finetuned Language Models Are Zero-Shot Learners

About

This paper explores a simple method for improving the zero-shot learning abilities of language models. We show that instruction tuning -- finetuning language models on a collection of tasks described via instructions -- substantially improves zero-shot performance on unseen tasks. We take a 137B parameter pretrained language model and instruction-tune it on over 60 NLP tasks verbalized via natural language instruction templates. We evaluate this instruction-tuned model, which we call FLAN, on unseen task types. FLAN substantially improves the performance of its unmodified counterpart and surpasses zero-shot 175B GPT-3 on 20 of 25 tasks that we evaluate. FLAN even outperforms few-shot GPT-3 by a large margin on ANLI, RTE, BoolQ, AI2-ARC, OpenbookQA, and StoryCloze. Ablation studies reveal that number of finetuning datasets, model scale, and natural language instructions are key to the success of instruction tuning.

Jason Wei, Maarten Bosma, Vincent Y. Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, Andrew M. Dai, Quoc V. Le• 2021

Related benchmarks

TaskDatasetResultRank
Question AnsweringOpenBookQA
Accuracy77.4
465
Natural Language InferenceRTE
Accuracy79.9
367
Instruction FollowingIFEval
Accuracy (0-100)75.9
292
Instruction FollowingAlpacaEval 2.0
LC Win Rate33.1
281
Reading ComprehensionBoolQ
Accuracy83.6
219
Natural Language InferenceSNLI
Accuracy62.3
174
General KnowledgeMMLU
MMLU General Knowledge Accuracy67.7
170
Mathematical Problem SolvingMATH
Accuracy51.7
166
Question AnsweringARC
Accuracy71
154
Natural Language InferenceMNLI (matched)
Accuracy60.8
110
Showing 10 of 68 rows

Other info

Code

Follow for update