Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

The Power of Scale for Parameter-Efficient Prompt Tuning

About

In this work, we explore "prompt tuning", a simple yet effective mechanism for learning "soft prompts" to condition frozen language models to perform specific downstream tasks. Unlike the discrete text prompts used by GPT-3, soft prompts are learned through backpropagation and can be tuned to incorporate signal from any number of labeled examples. Our end-to-end learned approach outperforms GPT-3's "few-shot" learning by a large margin. More remarkably, through ablations on model size using T5, we show that prompt tuning becomes more competitive with scale: as models exceed billions of parameters, our method "closes the gap" and matches the strong performance of model tuning (where all model weights are tuned). This finding is especially relevant in that large models are costly to share and serve, and the ability to reuse one frozen model for multiple downstream tasks can ease this burden. Our method can be seen as a simplification of the recently proposed "prefix tuning" of Li and Liang (2021), and we provide a comparison to this and other similar approaches. Finally, we show that conditioning a frozen model with soft prompts confers benefits in robustness to domain transfer, as compared to full model tuning.

Brian Lester, Rami Al-Rfou, Noah Constant• 2021

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K (test)
Accuracy72
900
Multi-task Language UnderstandingMMLU
Accuracy27.5
876
Image ClassificationFlowers102
Accuracy74.83
558
Natural Language UnderstandingGLUE
SST-294.61
531
Natural Language InferenceRTE
Accuracy77.13
448
Natural Language UnderstandingGLUE (test)
SST-2 Accuracy93.7
416
Question AnsweringSQuAD v1.1 (dev)
F1 Score12
380
Vision-and-Language NavigationR2R (val unseen)
Success Rate (SR)56.49
344
Subjectivity ClassificationSubj
Accuracy82.7
329
Commonsense ReasoningCommon Sense Reasoning Tasks
Avg Score36.96
316
Showing 10 of 242 rows
...

Other info

Follow for update