PLATO: Pre-trained Dialogue Generation Model with Discrete Latent Variable
About
Pre-training models have been proved effective for a wide range of natural language processing tasks. Inspired by this, we propose a novel dialogue generation pre-training framework to support various kinds of conversations, including chit-chat, knowledge grounded dialogues, and conversational question answering. In this framework, we adopt flexible attention mechanisms to fully leverage the bi-directional context and the uni-directional characteristic of language generation. We also introduce discrete latent variables to tackle the inherent one-to-many mapping problem in response generation. Two reciprocal tasks of response generation and latent act recognition are designed and carried out simultaneously within a shared network. Comprehensive experiments on three publicly available datasets verify the effectiveness and superiority of the proposed framework.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dialogue Generation | DailyDialog | Distinct-10.054 | 26 | |
| Dialogue Response Generation | Persona-Chat | BLEU-145.8 | 20 | |
| Response Generation | DailyDialog (test) | BLEU-232.2 | 16 | |
| Dialog Generation | DSTC7-AVSD (test) | BLEU-10.784 | 12 | |
| Dialog State Tracking | DuClarifyDial | Type Accuracy99 | 5 | |
| Response Generation | DuClarifyDial | BLEU-146 | 5 | |
| Dialog Act Planning | DuClarifyDial | Act Accuracy91 | 5 | |
| End-to-End Dialog Generation | DuClarifyDial (test) | BLEU-10.32 | 5 |