Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Wizard of Wikipedia: Knowledge-Powered Conversational agents

About

In open-domain dialogue intelligent agents should exhibit the use of knowledge, however there are few convincing demonstrations of this to date. The most popular sequence to sequence models typically "generate and hope" generic utterances that can be memorized in the weights of the model when mapping from input utterance(s) to output, rather than employing recalled knowledge as context. Use of knowledge has so far proved difficult, in part because of the lack of a supervised learning benchmark task which exhibits knowledgeable open dialogue with clear grounding. To that end we collect and release a large dataset with conversations directly grounded with knowledge retrieved from Wikipedia. We then design architectures capable of retrieving knowledge, reading and conditioning on it, and finally generating natural responses. Our best performing dialogue models are able to conduct knowledgeable discussions on open-domain topics as evaluated by automatic metrics and human evaluations, while our new benchmark allows for measuring further improvements in this important research direction.

Emily Dinan, Stephen Roller, Kurt Shuster, Angela Fan, Michael Auli, Jason Weston• 2018

Related benchmarks

TaskDatasetResultRank
Dialogue EvaluationIce-breaker human evaluation 1.0 (test)
Overall Score0.552
10
Open-domain Dialogue EvaluationFree Run 2 1.0 (secondary data collection run)
Overall Quality Score0.455
10
Open-domain Dialogue EvaluationFree run Mechanical Turk 1 (initial data collection run)
Overall Score0.534
10
Knowledge Grounded DialogueWizInternet 1.0 (test)
PPL22.3
10
Knowledge-Grounded ConversationPersonalized KGC dataset (test)
BLEU-11.05
9
Open-domain dialogueWizard-of-Wikipedia KILT (test)
F1 Score11.85
8
Knowledge Grounded DialogueWizards of Wikipedia
F1 Score35.5
6
Showing 7 of 7 rows

Other info

Follow for update