Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Prompting ELECTRA: Few-Shot Learning with Discriminative Pre-Trained Models

About

Pre-trained masked language models successfully perform few-shot learning by formulating downstream tasks as text infilling. However, as a strong alternative in full-shot settings, discriminative pre-trained models like ELECTRA do not fit into the paradigm. In this work, we adapt prompt-based few-shot learning to ELECTRA and show that it outperforms masked language models in a wide range of tasks. ELECTRA is pre-trained to distinguish if a token is generated or original. We naturally extend that to prompt-based few-shot learning by training to score the originality of the target options without introducing new parameters. Our method can be easily adapted to tasks involving multi-token predictions without extra computation overhead. Analysis shows that ELECTRA learns distributions that align better with downstream tasks.

Mengzhou Xia, Mikel Artetxe, Jingfei Du, Danqi Chen, Ves Stoyanov• 2022

Related benchmarks

TaskDatasetResultRank
Natural Language ProcessingT0 benchmark
RTE69
18
Showing 1 of 1 rows

Other info

Follow for update