Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ReGen: Zero-Shot Text Classification via Training Data Generation with Progressive Dense Retrieval

About

With the development of large language models (LLMs), zero-shot learning has attracted much attention for various NLP tasks. Different from prior works that generate training data with billion-scale natural language generation (NLG) models, we propose a retrieval-enhanced framework to create training data from a general-domain unlabeled corpus. To realize this, we first conduct contrastive pretraining to learn an unsupervised dense retriever for extracting the most relevant documents using class-descriptive verbalizers. We then further propose two simple strategies, namely Verbalizer Augmentation with Demonstrations and Self-consistency Guided Filtering to improve the topic coverage of the dataset while removing noisy examples. Experiments on nine datasets demonstrate that REGEN achieves 4.3% gain over the strongest baselines and saves around 70% of the time compared to baselines using large NLG models. Besides, REGEN can be naturally integrated with recently proposed large language models to boost performance.

Yue Yu, Yuchen Zhuang, Rongzhi Zhang, Yu Meng, Jiaming Shen, Chao Zhang• 2023

Related benchmarks

TaskDatasetResultRank
Sentiment ClassificationSST2 (test)
Accuracy87.84
214
Sentiment AnalysisSST-2
Accuracy85.32
156
Sentiment ClassificationIMDB (test)--
144
Topic ClassificationAG News (test)
Accuracy85
98
Topic ClassificationDBPedia (test)
Accuracy87.6
64
Sentiment AnalysisIMDB
Accuracy87.84
57
Sentiment ClassificationYelp (test)
Accuracy93
46
Topic ClassificationYahoo (test)
Accuracy59.4
36
Sentiment AnalysisYelp
Accuracy89
30
Sentiment AnalysisRotten Tomato
Accuracy81.42
25
Showing 10 of 13 rows

Other info

Code

Follow for update