Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Zero-Shot Text Classification with Self-Training

About

Recent advances in large pretrained language models have increased attention to zero-shot text classification. In particular, models finetuned on natural language inference datasets have been widely adopted as zero-shot classifiers due to their promising results and off-the-shelf availability. However, the fact that such models are unfamiliar with the target task can lead to instability and performance issues. We propose a plug-and-play method to bridge this gap using a simple self-training approach, requiring only the class names along with an unlabeled dataset, and without the need for domain expertise or trial and error. We show that fine-tuning the zero-shot classifier on its most confident predictions leads to significant performance gains across a wide range of text classification tasks, presumably since self-training adapts the zero-shot model to the task at hand.

Ariel Gera, Alon Halfon, Eyal Shnarch, Yotam Perlitz, Liat Ein-Dor, Noam Slonim• 2022

Related benchmarks

TaskDatasetResultRank
Topic ClassificationAG News (test)
Accuracy76.5
98
Topic ClassificationDBPedia (test)
Accuracy92.2
64
Sentiment ClassificationYelp (test)
Accuracy94.3
46
Topic ClassificationYahoo (test)
Accuracy59.8
36
Topic ClassificationNYT (test)
Accuracy92.5
18
Showing 5 of 5 rows

Other info

Follow for update