Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

PPT: Pre-trained Prompt Tuning for Few-shot Learning

About

Prompts for pre-trained language models (PLMs) have shown remarkable performance by bridging the gap between pre-training tasks and various downstream tasks. Among these methods, prompt tuning, which freezes PLMs and only tunes soft prompts, provides an efficient and effective solution for adapting large-scale PLMs to downstream tasks. However, prompt tuning is yet to be fully explored. In our pilot experiments, we find that prompt tuning performs comparably with conventional full-model fine-tuning when downstream data are sufficient, whereas it performs much worse under few-shot learning settings, which may hinder the application of prompt tuning in practice. We attribute this low performance to the manner of initializing soft prompts. Therefore, in this work, we propose to pre-train prompts by adding soft prompts into the pre-training stage to obtain a better initialization. We name this Pre-trained Prompt Tuning framework "PPT". To ensure the generalization of PPT, we formulate similar classification tasks into a unified task form and pre-train soft prompts for this unified task. Extensive experiments show that tuning pre-trained prompts for downstream tasks can reach or even outperform full-model fine-tuning under both full-data and few-shot settings. Our approach is effective and efficient for using large-scale PLMs in practice.

Yuxian Gu, Xu Han, Zhiyuan Liu, Minlie Huang• 2021

Related benchmarks

TaskDatasetResultRank
Sentiment ClassificationSST-2
Accuracy52.53
174
Text ClassificationYahoo! Answers (test)
Clean Accuracy70.5
133
Binary ClassificationGLUE (test)
QNLI Accuracy68.8
25
Natural Language InferenceRTE (val)
Accuracy0.918
24
Sentiment ClassificationCR (Entire dataset)
Accuracy64.03
24
Sentiment ClassificationMR
Accuracy50.5
24
Natural Language InferenceCB val (test)
Accuracy93.7
19
Sentiment ClassificationSST-2 32 samples
Accuracy94.4
11
Sentiment ClassificationSST-5 32 samples
Accuracy50.2
11
Text ClassificationTNEWS CLUE (test)
Accuracy50.6
11
Showing 10 of 18 rows

Other info

Code

Follow for update