Effectiveness of Pre-training for Few-shot Intent Classification
About
This paper investigates the effectiveness of pre-training for few-shot intent classification. While existing paradigms commonly further pre-train language models such as BERT on a vast amount of unlabeled corpus, we find it highly effective and efficient to simply fine-tune BERT with a small set of labeled utterances from public datasets. Specifically, fine-tuning BERT with roughly 1,000 labeled data yields a pre-trained model -- IntentBERT, which can easily surpass the performance of existing pre-trained models for few-shot intent classification on novel domains with very different semantics. The high effectiveness of IntentBERT confirms the feasibility and practicality of few-shot intent detection, and its high generalization ability across different domains suggests that intent classification tasks may share a similar underlying structure, which can be efficiently learned from a small set of labeled data. The source code can be found at https://github.com/hdzhang-code/IntentBERT.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Intent Classification | MCID 10-shot | Accuracy82.67 | 23 | |
| Intent Classification | HINT3 10-shot | Accuracy68.96 | 23 | |
| Intent Classification | HINT3 5-shot | Accuracy60.77 | 23 | |
| Intent Classification | HWU64 10-shot | Accuracy84.26 | 20 | |
| Intent Classification | BANKING77 10-shot | Accuracy83.94 | 20 | |
| Intent Classification | BANKING77 5-shot (test) | Accuracy70.64 | 20 | |
| Intent Classification | HWU64 10-shot (test) | Accuracy83.55 | 12 | |
| Intent Classification | HWU64 5-shot (test) | Accuracy77.6 | 12 | |
| Intent Classification | BANKING77 10-shot (test) | Accuracy81.18 | 12 | |
| Intent Classification | MCID 5-shot (test) | Accuracy0.7667 | 12 |