Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Conversations Are Not Flat: Modeling the Dynamic Information Flow across Dialogue Utterances

About

Nowadays, open-domain dialogue models can generate acceptable responses according to the historical context based on the large-scale pre-trained language models. However, they generally concatenate the dialogue history directly as the model input to predict the response, which we named as the flat pattern and ignores the dynamic information flow across dialogue utterances. In this work, we propose the DialoFlow model, in which we introduce a dynamic flow mechanism to model the context flow, and design three training objectives to capture the information dynamics across dialogue utterances by addressing the semantic influence brought about by each utterance in large-scale pre-training. Experiments on the multi-reference Reddit Dataset and DailyDialog Dataset demonstrate that our DialoFlow significantly outperforms the DialoGPT on the dialogue generation task. Besides, we propose the Flow score, an effective automatic metric for evaluating interactive human-bot conversation quality based on the pre-trained DialoFlow, which presents high chatbot-level correlation ($r=0.9$) with human ratings among 11 chatbots. Code and pre-trained models will be public. \footnote{\url{https://github.com/ictnlp/DialoFlow}}

Zekang Li, Jinchao Zhang, Zhengcong Fei, Yang Feng, Jie Zhou• 2021

Related benchmarks

TaskDatasetResultRank
Dialogue GenerationPersonaChat (test)--
27
Dialogue GenerationDailyDialog
Distinct-14.56
26
Response GenerationDailyDialog (test)
BLEU-228.02
16
Dialogue EvaluationDSTC9 Interactive Dialogue Evaluation Track (test)--
12
Response GenerationReddit multi-reference 6K (test)
NIST-23.9
9
Dialogue EvaluationDSTC9 Interactive Conversation (test)
Pearson0.91
3
Showing 6 of 6 rows

Other info

Code

Follow for update