Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

AttentionRetriever: Attention Layers are Secretly Long Document Retrievers

About

Retrieval augmented generation (RAG) has been widely adopted to help Large Language Models (LLMs) to process tasks involving long documents. However, existing retrieval models are not designed for long document retrieval and fail to address several key challenges of long document retrieval, including context-awareness, causal dependence, and scope of retrieval. In this paper, we proposed AttentionRetriever, a novel long document retrieval model that leverages attention mechanism and entity-based retrieval to build context-aware embeddings for long document and determine the scope of retrieval. With extensive experiments, we found AttentionRetriever is able to outperform existing retrieval models on long document retrieval datasets by a large margin while remaining as efficient as dense retrieval models.

David Jiahao Fu, Lam Thanh Do, Jiayu Li, Kevin Chen-Chuan Chang• 2026

Related benchmarks

TaskDatasetResultRank
Long document retrievalLongBench Retrieval v2 (full)
F1 Score0.4843
55
Question AnsweringMultifieldQA
F1 Score54.36
52
Single-document retrievalQASA
F1 Score57.74
44
Single-document retrievalQasper
F1 Score50.3
44
Single-document retrievalRepLiQA
F1 Score0.8555
44
Single-document retrievalConditionalQA
F135.28
44
Single-document retrievalNaturalQuestions
F1 Score60.65
44
Question AnsweringQasper
F1 Score0.322
16
Question AnsweringNarrativeQA
F1 Score23.93
16
Long document retrievalQASA (test)
F1 Score56.63
11
Showing 10 of 24 rows

Other info

Follow for update