Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

G-reasoner: Foundation Models for Unified Reasoning over Graph-structured Knowledge

About

Large language models (LLMs) excel at complex reasoning but remain limited by static and incomplete parametric knowledge. Retrieval-augmented generation (RAG) mitigates this by incorporating external knowledge, yet existing RAGs struggle with knowledge-intensive tasks due to fragmented information and weak modeling of knowledge structure. Graphs offer a natural way to model relationships within knowledge, but LLMs are inherently unstructured and cannot effectively reason over graph-structured data. Recent graph-enhanced RAG (GraphRAG) attempts to bridge this gap by constructing tailored graphs and enabling LLMs to reason on them. However, these methods often depend on ad-hoc graph designs, heuristic search, or costly agent pipelines, which hinder scalability and generalization. To address these challenges, we present G-reasoner, a unified framework that integrates graph and language foundation models for scalable reasoning over diverse graph-structured knowledge. Central to our approach is QuadGraph, a standardized four-layer abstraction that unifies heterogeneous knowledge sources into a common graph representation. Building on this, we introduce a 34M-parameter graph foundation model (GFM) that jointly captures graph topology and textual semantics, and is integrated with LLMs to enhance reasoning in downstream applications. To ensure scalability and efficiency, mixed-precision training and distributed message-passing are implemented to scale GFM with more GPUs. Extensive experiments on six benchmarks show that G-reasoner consistently outperforms state-of-the-art baselines, significantly enhances LLM reasoning, and achieves strong efficiency and cross-graph generalization.

Linhao Luo, Zicheng Zhao, Junnan Liu, Zhangchi Qiu, Junnan Dong, Serge Panev, Chen Gong, Thuy-Trang Vu, Gholamreza Haffari, Dinh Phung, Alan Wee-Chung Liew, Shirui Pan• 2025

Related benchmarks

TaskDatasetResultRank
Multi-hop Question Answering2Wiki
Exact Match74.9
152
Multi-hop QA RetrievalMuSiQue
R@254.8
36
Multi-hop RetrievalHotpotQA
Recall@285.9
23
Multi-hop QA Retrieval2Wiki
Recall@281.2
23
Question AnsweringG-bench Novel
Accuracy58.9
11
Question AnsweringG-bench Medical
Accuracy73.3
11
Question AnsweringG-bench CS
Accuracy73.9
11
Evidence RetrievalG-bench Novel
Recall87.7
10
Evidence RetrievalG-bench Medical
Recall93.8
10
Graph ReasoningG-bench CS
Inference Time (s)0.2
9
Showing 10 of 14 rows

Other info

Follow for update