MinTL: Minimalist Transfer Learning for Task-Oriented Dialogue Systems
About
In this paper, we propose Minimalist Transfer Learning (MinTL) to simplify the system design process of task-oriented dialogue systems and alleviate the over-dependency on annotated data. MinTL is a simple yet effective transfer learning framework, which allows us to plug-and-play pre-trained seq2seq models, and jointly learn dialogue state tracking and dialogue response generation. Unlike previous approaches, which use a copy mechanism to "carryover" the old dialogue states to the new one, we introduce Levenshtein belief spans (Lev), that allows efficient dialogue state tracking with a minimal generation length. We instantiate our learning framework with two pre-trained backbones: T5 and BART, and evaluate them on MultiWOZ. Extensive experiments demonstrate that: 1) our systems establish new state-of-the-art results on end-to-end response generation, 2) MinTL-based systems are more robust than baseline methods in the low resource setting, and they achieve competitive results with only 20\% training data, and 3) Lev greatly improves the inference efficiency.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Dialog State Tracking | MultiWOZ 2.1 (test) | Joint Goal Accuracy53.62 | 88 | |
| Dialogue State Tracking | MultiWOZ 2.1 (test) | Joint Goal Accuracy53.62 | 85 | |
| End-to-end task-oriented dialogue | MultiWOZ (test) | Task Success Rate74.9 | 68 | |
| Dialog State Tracking | MultiWOZ 2.0 (test) | Joint Goal Accuracy52.1 | 47 | |
| Task-oriented Dialogue | MultiWOZ 2.0 (test) | Inform Rate84.88 | 37 | |
| Task-oriented Dialogue | MultiWOZ 2.2 (test) | Inform Rate73.7 | 23 | |
| End-to-end Dialogue Modelling | MultiWOZ 2.0 (test) | Inform Rate84.88 | 22 | |
| End-to-end task-oriented dialogue | MultiWOZ 2.0 (test) | Inform Accuracy84.88 | 22 | |
| Task-oriented Dialogue | MultiWOZ 5% 2.0 (train) | Inform75.48 | 10 | |
| Task-oriented Dialogue | MultiWOZ 2.0 (10% train) | Inform Rate78.08 | 10 |