Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Pre-trained Language Models Can be Fully Zero-Shot Learners

About

How can we extend a pre-trained model to many language understanding tasks, without labeled or additional unlabeled data? Pre-trained language models (PLMs) have been effective for a wide range of NLP tasks. However, existing approaches either require fine-tuning on downstream labeled datasets or manually constructing proper prompts. In this paper, we propose nonparametric prompting PLM (NPPrompt) for fully zero-shot language understanding. Unlike previous methods, NPPrompt uses only pre-trained language models and does not require any labeled data or additional raw corpus for further fine-tuning, nor does it rely on humans to construct a comprehensive set of prompt label words. We evaluate NPPrompt against previous major few-shot and zero-shot learning methods on diverse NLP tasks: including text classification, text entailment, similar text retrieval, and paraphrasing. Experimental results demonstrate that our NPPrompt outperforms the previous best fully zero-shot method by big margins, with absolute gains of 12.8% in accuracy on text classification and 18.9% on the GLUE benchmark.

Xuandong Zhao, Siqi Ouyang, Zhiguo Yu, Ming Wu, Lei Li• 2022

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE
SST-286.3
452
Text ClassificationAGNews
Accuracy85.2
119
Text ClassificationDBpedia (DBP)
Accuracy86.8
110
Text ClassificationIMDB
Accuracy94.2
107
Multiple-choice Question AnsweringCommonsenseQA (dev)
Accuracy34.2
18
Showing 5 of 5 rows

Other info

Code

Follow for update