Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Think Parallax: Solving Multi-Hop Problems via Multi-View Knowledge-Graph-Based Retrieval-Augmented Generation

About

Large language models (LLMs) still struggle with multi-hop reasoning over knowledge-graphs (KGs), and we identify a previously overlooked structural reason for this difficulty: Transformer attention heads naturally specialize in distinct semantic relations across reasoning stages, forming a hop-aligned relay pattern. This key finding suggests that multi-hop reasoning is inherently multi-view, yet existing KG-based retrieval-augmented generation (KG-RAG) systems collapse all reasoning hops into a single representation, flat embedding space, suppressing this implicit structure and causing noisy or drifted path exploration. We introduce ParallaxRAG, a symmetric multi-view framework that decouples queries and KGs into aligned, head-specific semantic spaces. By enforcing relational diversity across multiple heads while constraining weakly related paths, ParallaxRAG constructs more accurate, cleaner subgraphs and guides LLMs through grounded, hop-wise reasoning. On WebQSP and CWQ, it achieves state-of-the-art retrieval and QA performance, substantially reduces hallucination, and generalizes strongly to the biomedical BioASQ benchmark.

Jinliang Liu, Jiale Bai, Shaoning Zeng• 2025

Related benchmarks

TaskDatasetResultRank
Knowledge Graph Question AnsweringCWQ
Hit@170.74
166
Knowledge Graph Question AnsweringWebQSP
Hit Rate0.9353
19
Knowledge Graph RetrievalWEBQSP (test)
Triple Recall (1-hop)96.6
9
Knowledge Graph RetrievalCWQ (Complex WebQuestions) (test)
Triple Recall (1-hop) (SP)91.6
7
List Question AnsweringBioASQ Task B
Precision71.16
6
Yes/No Question AnsweringBioASQ Task B
Accuracy93.51
6
Factoid Question AnsweringBioASQ Task B
Strict Score42.1
6
Showing 7 of 7 rows

Other info

Follow for update