Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Pre-training Intent-Aware Encoders for Zero- and Few-Shot Intent Classification

About

Intent classification (IC) plays an important role in task-oriented dialogue systems. However, IC models often generalize poorly when training without sufficient annotated examples for each user intent. We propose a novel pre-training method for text encoders that uses contrastive learning with intent psuedo-labels to produce embeddings that are well-suited for IC tasks, reducing the need for manual annotations. By applying this pre-training strategy, we also introduce Pre-trained Intent-aware Encoder (PIE), which is designed to align encodings of utterances with their intent names. Specifically, we first train a tagger to identify key phrases within utterances that are crucial for interpreting intents. We then use these extracted phrases to create examples for pre-training a text encoder in a contrastive manner. As a result, our PIE model achieves up to 5.4% and 4.0% higher accuracy than the previous state-of-the-art text encoder for the N-way zero- and one-shot settings on four IC datasets.

Mujeen Sung, James Gung, Elman Mansimov, Nikolaos Pappas, Raphael Shu, Salvatore Romeo, Yi Zhang, Vittorio Castelli• 2023

Related benchmarks

TaskDatasetResultRank
Triplet Evaluation (Ori-Imp)Intent Semantic Toolkit (test)
Teasy Score67.3
7
Triplet TaskIntent Semantic Toolkit Ori-Ori
Thard24
4
Intent ClusteringIntent Semantic Toolkit
Agg Score59.9
4
Binary Intent ClassificationIntent Semantic Toolkit
Ori Accuracy86.6
4
Intent ClusteringOriginal Dataset
KM Score83.4
4
Multi-class Intent ClassificationOriginal Dataset
10-shot Accuracy84.7
4
Multi-class Intent ClassificationIntent Semantic Toolkit
Accuracy (0-shot)25.4
4
Showing 7 of 7 rows

Other info

Follow for update