Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

FactCorrector: A Graph-Inspired Approach to Long-Form Factuality Correction of Large Language Models

About

Large language models (LLMs) are widely used in knowledge-intensive applications but often generate factually incorrect responses. A promising approach to rectify these flaws is correcting LLMs using feedback. Therefore, in this paper, we introduce FactCorrector, a new post-hoc correction method that adapts across domains without retraining and leverages structured feedback about the factuality of the original response to generate a correction. To support rigorous evaluations of factuality correction methods, we also develop the VELI5 benchmark, a novel dataset containing systematically injected factual errors and ground-truth corrections. Experiments on VELI5 and several popular long-form factuality datasets show that the FactCorrector approach significantly improves factual precision while preserving relevance, outperforming strong baselines. We release our code at https://ibm.biz/factcorrector.

Javier Carnerero-Cano, Massimiliano Pronesti, Radu Marinescu, Tigran Tchrakian, James Barry, Jasmina Gajcin, Yufang Hou, Alessandra Pascale, Elizabeth Daly• 2026

Related benchmarks

TaskDatasetResultRank
Factuality CorrectionVELI5
Mean Factual Precision0.97
64
Factuality CorrectionBio (test)
Precision46
44
Factuality CorrectionASKHIST
Mean Factual Precision0.92
40
Factual CorrectionCONFLICTS
ROUGE97
25
Factuality CorrectionVELI5 1.0 (test)
Precision (Pr)36
24
Factuality CorrectionBIO dataset
Factual Precision91
24
Post-hoc CorrectionConflictBank (100 atomic claims)
ROUGE89
15
Showing 7 of 7 rows

Other info

Follow for update