Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Effectiveness of Pre-training for Few-shot Intent Classification

About

This paper investigates the effectiveness of pre-training for few-shot intent classification. While existing paradigms commonly further pre-train language models such as BERT on a vast amount of unlabeled corpus, we find it highly effective and efficient to simply fine-tune BERT with a small set of labeled utterances from public datasets. Specifically, fine-tuning BERT with roughly 1,000 labeled data yields a pre-trained model -- IntentBERT, which can easily surpass the performance of existing pre-trained models for few-shot intent classification on novel domains with very different semantics. The high effectiveness of IntentBERT confirms the feasibility and practicality of few-shot intent detection, and its high generalization ability across different domains suggests that intent classification tasks may share a similar underlying structure, which can be efficiently learned from a small set of labeled data. The source code can be found at https://github.com/hdzhang-code/IntentBERT.

Haode Zhang, Yuwei Zhang, Li-Ming Zhan, Jiaxin Chen, Guangyuan Shi, Albert Y.S. Lam, Xiao-Ming Wu• 2021

Related benchmarks

TaskDatasetResultRank
Intent ClassificationMCID 10-shot
Accuracy82.67
23
Intent ClassificationHINT3 10-shot
Accuracy68.96
23
Intent ClassificationHINT3 5-shot
Accuracy60.77
23
Intent ClassificationHWU64 10-shot
Accuracy84.26
20
Intent ClassificationBANKING77 10-shot
Accuracy83.94
20
Intent ClassificationBANKING77 5-shot (test)
Accuracy70.64
20
Intent ClassificationHWU64 10-shot (test)
Accuracy83.55
12
Intent ClassificationHWU64 5-shot (test)
Accuracy77.6
12
Intent ClassificationBANKING77 10-shot (test)
Accuracy81.18
12
Intent ClassificationMCID 5-shot (test)
Accuracy0.7667
12
Showing 10 of 13 rows

Other info

Follow for update