TDGNet: Hallucination Detection in Diffusion Language Models via Temporal Dynamic Graphs
About
Diffusion language models (D-LLMs) offer parallel denoising and bidirectional context, but hallucination detection for D-LLMs remains underexplored. Prior detectors developed for auto-regressive LLMs typically rely on single-pass cues and do not directly transfer to diffusion generation, where factuality evidence is distributed across the denoising trajectory and may appear, drift, or be self-corrected over time. We introduce TDGNet, a temporal dynamic graph framework that formulates hallucination detection as learning over evolving token-level attention graphs. At each denoising step, we sparsify the attention graph and update per-token memories via message passing, then apply temporal attention to aggregate trajectory-wide evidence for final prediction. Experiments on LLaDA-8B and Dream-7B across QA benchmarks show consistent AUROC improvements over output-based, latent-based, and static-graph baselines, with single-pass inference and modest overhead. These results highlight the importance of temporal reasoning on attention graphs for robust hallucination detection in diffusion language models.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Hallucination Detection | TriviaQA | AUROC0.74 | 265 | |
| Hallucination Detection | TriviaQA (test) | AUC-ROC85 | 169 | |
| Hallucination Detection | HotpotQA | AUROC0.74 | 118 | |
| Hallucination Detection | NQ (test) | AUC ROC87 | 84 | |
| Hallucination Detection | MATH | Mean AUROC72 | 72 | |
| Hallucination Detection | CSQA | AUROC72 | 55 |