Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Enhancing Uncertainty-Based Hallucination Detection with Stronger Focus

About

Large Language Models (LLMs) have gained significant popularity for their impressive performance across diverse fields. However, LLMs are prone to hallucinate untruthful or nonsensical outputs that fail to meet user expectations in many real-world applications. Existing works for detecting hallucinations in LLMs either rely on external knowledge for reference retrieval or require sampling multiple responses from the LLM for consistency verification, making these methods costly and inefficient. In this paper, we propose a novel reference-free, uncertainty-based method for detecting hallucinations in LLMs. Our approach imitates human focus in factuality checking from three aspects: 1) focus on the most informative and important keywords in the given text; 2) focus on the unreliable tokens in historical context which may lead to a cascade of hallucinations; and 3) focus on the token properties such as token type and token frequency. Experimental results on relevant datasets demonstrate the effectiveness of our proposed method, which achieves state-of-the-art performance across all the evaluation metrics and eliminates the need for additional information.

Tianhang Zhang, Lin Qiu, Qipeng Guo, Cheng Deng, Yue Zhang, Zheng Zhang, Chenghu Zhou, Xinbing Wang, Luoyi Fu• 2023

Related benchmarks

TaskDatasetResultRank
Hallucination DetectionTriviaQA
AUROC0.589
265
Hallucination DetectionTriviaQA (test)
AUC-ROC81.7
169
Hallucination DetectionHaluEval (test)
AUC-ROC78.1
126
Hallucination DetectionTruthfulQA (test)
AUC-ROC81.4
91
Hallucination DetectionHELM Sentence Level v1.0 (test)
AUC0.7593
84
Hallucination DetectionHELM Passage Level v1.0 (test)
AUC0.8659
84
Hallucination DetectionNQ (test)
AUC ROC79.4
84
Hallucination DetectionCompany
AUC-ROC0.556
68
Hallucination DetectionVQA v2
ROC0.623
27
Hallucination DetectionHalLoc VQA
ROC65
27
Showing 10 of 25 rows

Other info

Follow for update