Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

MIDAS: Multi-level Intent, Domain, And Slot Knowledge Distillation for Multi-turn NLU

About

Although Large Language Models (LLMs) can generate coherent text, they often struggle to recognise user intent behind queries. In contrast, Natural Language Understanding (NLU) models interpret the purpose and key information of user input for responsive interactions. Existing NLU models typically map utterances to a dual-level semantic frame, involving sentence-level intent (SI) and word-level slot (WS) labels. However, real-life conversations primarily consist of multi-turn dialogues, requiring the interpretation of complex and extended exchanges. Researchers encounter challenges in addressing all facets of multi-turn dialogue using a unified NLU model. This paper introduces MIDAS, a novel approach leveraging multi-level intent, domain, and slot knowledge distillation for multi-turn NLU. We construct distinct teachers for SI detection, WS filling, and conversation-level domain (CD) classification, each fine-tuned for specific knowledge. A multi-teacher loss is proposed to facilitate the integration of these teachers, guiding a student model in multi-turn dialogue tasks. Results demonstrate the efficacy of our model in improving multi-turn conversation understanding, showcasing the potential for advancements in NLU through multi-level dialogue knowledge distillation. Our implementation is open-sourced on https://github.com/adlnlp/Midas.

Yan Li, So-Eon Kim, Seong-Bae Park, Soyeon Caren Han• 2024

Related benchmarks

TaskDatasetResultRank
Intent DetectionMWOZ
Accuracy85.02
18
Intent DetectionM2M
Accuracy94.27
18
Slot FillingMWOZ
Micro F199.28
18
Slot FillingM2M
Micro F198.56
18
Domain ClassificationMWOZ
Accuracy88.16
16
Domain ClassificationM2M
Accuracy89.52
16
Showing 6 of 6 rows

Other info

Code

Follow for update