Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Semantically Conditioned LSTM-based Natural Language Generation for Spoken Dialogue Systems

About

Natural language generation (NLG) is a critical component of spoken dialogue and it has a significant impact both on usability and perceived quality. Most NLG systems in common use employ rules and heuristics and tend to generate rigid and stylised responses without the natural variation of human language. They are also not easily scaled to systems covering multiple domains and languages. This paper presents a statistical language generator based on a semantically controlled Long Short-term Memory (LSTM) structure. The LSTM generator can learn from unaligned data by jointly optimising sentence planning and surface realisation using a simple cross entropy training criterion, and language variation can be easily achieved by sampling from output candidates. With fewer heuristics, an objective evaluation in two differing test domains showed the proposed method improved performance compared to previous methods. Human judges scored the LSTM system higher on informativeness and naturalness and overall preferred it to the other systems.

Tsung-Hsien Wen, Milica Gasic, Nikola Mrksic, Pei-Hao Su, David Vandyke, Steve Young• 2015

Related benchmarks

TaskDatasetResultRank
Natural language generationE2E (test)--
79
Response GenerationMultiWOZ (test)
BLEU Score21.6
27
Natural language generationTV RNNLG benchmark (test)
BLEU Score0.5265
4
Natural language generationLaptop RNNLG benchmark (test)
BLEU51.16
4
Dialog Response GenerationFEWSHOTWOZ (test)
Informativeness2.29
4
Dialog Response GenerationFEWSHOTWOZ Restaurant 1.0 (test)
BLEU15.9
3
Dialog Response GenerationFEWSHOTWOZ Laptop 1.0 (test)
BLEU21.98
3
Dialog Response GenerationFEWSHOTWOZ Hotel 1.0 (test)
BLEU31.3
3
Dialog Response GenerationFEWSHOTWOZ TV 1.0 (test)
BLEU22.39
3
Dialog Response GenerationFEWSHOTWOZ Attraction 1.0 (test)
BLEU7.76
3
Showing 10 of 13 rows

Other info

Follow for update