Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

TDGNet: Hallucination Detection in Diffusion Language Models via Temporal Dynamic Graphs

About

Diffusion language models (D-LLMs) offer parallel denoising and bidirectional context, but hallucination detection for D-LLMs remains underexplored. Prior detectors developed for auto-regressive LLMs typically rely on single-pass cues and do not directly transfer to diffusion generation, where factuality evidence is distributed across the denoising trajectory and may appear, drift, or be self-corrected over time. We introduce TDGNet, a temporal dynamic graph framework that formulates hallucination detection as learning over evolving token-level attention graphs. At each denoising step, we sparsify the attention graph and update per-token memories via message passing, then apply temporal attention to aggregate trajectory-wide evidence for final prediction. Experiments on LLaDA-8B and Dream-7B across QA benchmarks show consistent AUROC improvements over output-based, latent-based, and static-graph baselines, with single-pass inference and modest overhead. These results highlight the importance of temporal reasoning on attention graphs for robust hallucination detection in diffusion language models.

Arshia Hemmat, Philip Torr, Yongqiang Chen, Junchi Yu• 2026

Related benchmarks

TaskDatasetResultRank
Hallucination DetectionTriviaQA
AUROC0.74
265
Hallucination DetectionTriviaQA (test)
AUC-ROC85
169
Hallucination DetectionHotpotQA
AUROC0.74
118
Hallucination DetectionNQ (test)
AUC ROC87
84
Hallucination DetectionMATH
Mean AUROC72
72
Hallucination DetectionCSQA
AUROC72
55
Showing 6 of 6 rows

Other info

Follow for update