BERT for Joint Intent Classification and Slot Filling
About
Intent classification and slot filling are two essential tasks for natural language understanding. They often suffer from small-scale human-labeled training data, resulting in poor generalization capability, especially for rare words. Recently a new language representation model, BERT (Bidirectional Encoder Representations from Transformers), facilitates pre-training deep bidirectional representations on large-scale unlabeled corpora, and has created state-of-the-art models for a wide variety of natural language processing tasks after simple fine-tuning. However, there has not been much effort on exploring BERT for natural language understanding. In this work, we propose a joint intent classification and slot filling model based on BERT. Experimental results demonstrate that our proposed model achieves significant improvement on intent classification accuracy, slot filling F1, and sentence-level semantic frame accuracy on several public benchmark datasets, compared to the attention-based recurrent neural network models and slot-gated models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Joint Multiple Intent Detection and Slot Filling | MixSNIPS (test) | Slot F195.9 | 57 | |
| Joint Multiple Intent Detection and Slot Filling | MixATIS (test) | F1 Score (Slot)86.3 | 42 | |
| Intent Classification | Snips (test) | Accuracy98.6 | 40 | |
| Natural Language Understanding | Snips (test) | Intent Acc98.6 | 27 | |
| Intent Detection | ATIS | ID Accuracy97.5 | 27 | |
| Slot Filling | Snips (test) | F1 Score0.97 | 25 | |
| Spoken Language Understanding | ATIS (test) | Slot F196.1 | 18 | |
| Spoken Language Understanding | SNIPS | Slot F197 | 15 | |
| Slot Filling | ATIS | F1 Score96.1 | 14 | |
| Natural Language Understanding | ATIS (test) | Intent Accuracy97.9 | 12 |