Entailment as Few-Shot Learner
About
Large pre-trained language models (LMs) have demonstrated remarkable ability as few-shot learners. However, their success hinges largely on scaling model parameters to a degree that makes it challenging to train and serve. In this paper, we propose a new approach, named as EFL, that can turn small LMs into better few-shot learners. The key idea of this approach is to reformulate potential NLP task into an entailment one, and then fine-tune the model with as little as 8 examples. We further demonstrate our proposed method can be: (i) naturally combined with an unsupervised contrastive learning-based data augmentation method; (ii) easily extended to multilingual few-shot learning. A systematic evaluation on 18 standard NLP tasks demonstrates that this approach improves the various existing SOTA few-shot learning methods by 12\%, and yields competitive few-shot performance with 500 times larger models, such as GPT-3.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Natural Language Inference | SNLI (test) | Accuracy93.1 | 681 | |
| Text Classification | SST-2 (test) | Accuracy96.9 | 185 | |
| Natural Language Inference | SNLI | Accuracy93.1 | 174 | |
| Natural Language Inference | SNLI (dev) | Accuracy93.1 | 71 | |
| Binary Classification | GLUE (test) | QNLI Accuracy68 | 25 | |
| Sentiment Analysis | SST5 | F1 Score50.8 | 16 | |
| Sentiment Analysis | Amazon Reviews | F1 Score58.6 | 16 | |
| Sentence Classification | MPQA (test) | Accuracy90.8 | 15 | |
| Natural Language Inference | RTE (dev) | Accuracy90.5 | 12 | |
| Sentence Classification | MR full (test) | Accuracy92.5 | 9 |