Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Beyond Goldfish Memory: Long-Term Open-Domain Conversation

About

Despite recent improvements in open-domain dialogue models, state of the art models are trained and evaluated on short conversations with little context. In contrast, the long-term conversation setting has hardly been studied. In this work we collect and release a human-human dataset consisting of multiple chat sessions whereby the speaking partners learn about each other's interests and discuss the things they have learnt from past sessions. We show how existing models trained on existing datasets perform poorly in this long-term conversation setting in both automatic and human evaluations, and we study long-context models that can perform much better. In particular, we find retrieval-augmented methods and methods with an ability to summarize and recall previous conversations outperform the standard encoder-decoder architectures currently considered state of the art.

Jing Xu, Arthur Szlam, Jason Weston• 2021

Related benchmarks

TaskDatasetResultRank
Dialogue Response GenerationKEEM (KMSC memories) 1.0 (test)
Perplexity9.51
14
Dialogue ModelingMSC (Multi-Session Chat) 1.0 (test)
Session 1 Perplexity8.2
10
Language ModelingMSC Session 2 1.0 (val)
Perplexity9.08
10
Language ModelingMSC Session 3 1.0 (val)
Perplexity8.96
10
Language ModelingMSC Session 4 1.0 (val)
Perplexity9.07
10
Language ModelingMSC Session 5 1.0 (val)
Perplexity8.99
10
Language ModelingMSC Session Openings 1.0 (val)
Perplexity7.78
10
Language ModelingMSC Session 1 1.0 (val)
Perplexity8.16
10
Dialogue Response GenerationKEEM memories 1.0 (test)
Perplexity7.88
7
Multi-Session Dialogue GenerationMSC (Multi-Session Chat) Human Evaluation 1.0 (test)
Reference Own Topic24.2
6
Showing 10 of 10 rows

Other info

Code

Follow for update