Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts

About

Current Large Language Models (LLMs) are not only limited to some maximum context length, but also are not able to robustly consume long inputs. To address these limitations, we propose ReadAgent, an LLM agent system that increases effective context length up to 20x in our experiments. Inspired by how humans interactively read long documents, we implement ReadAgent as a simple prompting system that uses the advanced language capabilities of LLMs to (1) decide what content to store together in a memory episode, (2) compress those memory episodes into short episodic memories called gist memories, and (3) take actions to look up passages in the original text if ReadAgent needs to remind itself of relevant details to complete a task. We evaluate ReadAgent against baselines using retrieval methods, using the original long contexts, and using the gist memories. These evaluations are performed on three long-document reading comprehension tasks: QuALITY, NarrativeQA, and QMSum. ReadAgent outperforms the baselines on all three tasks while extending the effective context window by 3.5-20x.

Kuang-Huei Lee, Xinyun Chen, Hiroki Furuta, John Canny, Ian Fischer• 2024

Related benchmarks

TaskDatasetResultRank
Multi-hop Question AnsweringLocomo
F114.61
67
Long-context Question AnsweringLocomo
Average F125.87
64
Open-domain Question AnsweringLocomo
F10.0884
53
Single-hop Question AnsweringLocomo
F10.1246
53
Question AnsweringMuSiQue (test)
F1 Score45.1
43
Long-context reasoning and retrievalLoCoMo (test)
Single-Hop F185.7
37
Temporal Question AnsweringLocomo
F10.126
36
GUI NavigationMultimodal-Mind2Web Cross-Website
Step Success Rate31.1
32
GUI NavigationMultimodal-Mind2Web Cross-Domain
Step Success Rate33.4
27
GUI NavigationMultimodal-Mind2Web Cross-Task
Step Success Rate29.2
27
Showing 10 of 48 rows

Other info

Follow for update