Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Learning Dialogue Representations from Consecutive Utterances

About

Learning high-quality dialogue representations is essential for solving a variety of dialogue-oriented tasks, especially considering that dialogue systems often suffer from data scarcity. In this paper, we introduce Dialogue Sentence Embedding (DSE), a self-supervised contrastive learning method that learns effective dialogue representations suitable for a wide range of dialogue tasks. DSE learns from dialogues by taking consecutive utterances of the same dialogue as positive pairs for contrastive learning. Despite its simplicity, DSE achieves significantly better representation capability than other dialogue representation and universal sentence representation models. We evaluate DSE on five downstream dialogue tasks that examine dialogue representation at different semantic granularities. Experiments in few-shot and zero-shot settings show that DSE outperforms baselines by a large margin. For example, it achieves 13% average performance improvement over the strongest unsupervised baseline in 1-shot intent classification on 6 datasets. We also provide analyses on the benefits and limitations of our model.

Zhihan Zhou, Dejiao Zhang, Wei Xiao, Nicholas Dingwall, Xiaofei Ma, Andrew O. Arnold, Bing Xiang• 2022

Related benchmarks

TaskDatasetResultRank
Short Text ClusteringTweet
Accuracy50.8
28
Intent RecognitionOOS (test)
Overall Accuracy84.3
19
ClusteringBank77
NMI71
19
Response SelectionMWOZ 2.1
Accuracy (1/100)63.3
17
ClusteringCLINC
Accuracy62.1
15
Dialogue State TrackingMultiWOZ 2.1 (5%)
Joint Goal Acc23.8
11
Dialogue State TrackingMultiWOZ 2.1 (1%)
Joint Goal Acc9.8
10
Dialogue act predictionMWOZ (Full Data)
Micro-F191.7
7
Dialogue act predictionDSTC2
Micro-F1 Score92.6
7
Dialogue State TrackingMWOZ 2.1 (Full Data)
Joint Accuracy49.9
6
Showing 10 of 16 rows

Other info

Follow for update