Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Entropy Alone is Insufficient for Safe Selective Prediction in LLMs

About

Selective prediction systems can mitigate harms resulting from language model hallucinations by abstaining from answering in high-risk cases. Uncertainty quantification techniques are often employed to identify such cases, but are rarely evaluated in the context of the wider selective prediction policy and its ability to operate at low target error rates. We identify a model-dependent failure mode of entropy-based uncertainty methods that leads to unreliable abstention behaviour, and address it by combining entropy scores with a correctness probe signal. We find that across three QA benchmarks (TriviaQA, BioASQ, MedicalQA) and four model families, the combined score generally improves both the risk--coverage trade-off and calibration performance relative to entropy-only baselines. Our results highlight the importance of deployment-facing evaluation of uncertainty methods, using metrics that directly reflect whether a system can be trusted to operate at a stated risk level.

Edward Phillips, Fredrik K. Gustafsson, Sean Wu, Anshul Thakur, David A. Clifton• 2026

Related benchmarks

TaskDatasetResultRank
Hallucination DetectionTriviaQA
AUROC0.9102
438
Hallucination DetectionMedicalQA
AUROC78.95
28
Selective PredictionMedicalQA
E-AURC0.3053
28
Hallucination DetectionBioASQ
AUROC79.13
28
Selective PredictionTriviaQA
E-AURC0.1583
28
Selective PredictionBioASQ
E-AURC0.2173
28
Showing 6 of 6 rows

Other info

Follow for update