Multi-Task Pre-Training for Plug-and-Play Task-Oriented Dialogue System
About
Pre-trained language models have been recently shown to benefit task-oriented dialogue (TOD) systems. Despite their success, existing methods often formulate this task as a cascaded generation problem which can lead to error accumulation across different sub-tasks and greater data annotation overhead. In this study, we present PPTOD, a unified plug-and-play model for task-oriented dialogue. In addition, we introduce a new dialogue multi-task pre-training strategy that allows the model to learn the primary TOD task completion skills from heterogeneous dialog corpora. We extensively test our model on three benchmark TOD tasks, including end-to-end dialogue modelling, dialogue state tracking, and intent classification. Experimental results show that PPTOD achieves new state of the art on all evaluated tasks in both high-resource and low-resource scenarios. Furthermore, comparisons against previous SOTA methods show that the responses generated by PPTOD are more factually correct and semantically coherent as judged by human annotators.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dialog State Tracking | MultiWOZ 2.1 (test) | Joint Goal Accuracy57.45 | 88 | |
| Dialogue State Tracking | MultiWOZ 2.1 (test) | Joint Goal Accuracy57.45 | 85 | |
| End-to-end task-oriented dialogue | MultiWOZ 2.1 (test) | BLEU Score19.17 | 49 | |
| Dialog State Tracking | MultiWOZ 2.0 (test) | Joint Goal Accuracy53.89 | 47 | |
| Task-oriented Dialogue | MultiWOZ 2.0 (test) | Inform Rate89.2 | 37 | |
| End-to-end Dialogue Modelling | MultiWOZ 2.0 (test) | Inform Rate89.2 | 22 | |
| Task-oriented Dialogue Response Generation | Multi-WOZ 2.1 (test) | BLEU19.17 | 22 | |
| End-to-end task-oriented dialogue | MultiWOZ 2.0 (test) | Inform Accuracy89.2 | 22 | |
| Dialogue State Tracking | MultiWOZ 2.0 (test) | Joint Goal Accuracy53.89 | 13 | |
| Task-oriented Dialogue | MultiWOZ 2.1 (test) | Inform Rate87.09 | 11 |