Evolutionary Search for Automated Design of Uncertainty Quantification Methods
About
Uncertainty quantification (UQ) methods for large language models are predominantly designed by hand based on domain knowledge and heuristics, limiting their scalability and generality. We apply LLM-powered evolutionary search to automatically discover unsupervised UQ methods represented as Python programs. On the task of atomic claim verification, our evolved methods outperform strong manually-designed baselines, achieving up to 6.7% relative ROC-AUC improvement across 9 datasets while generalizing robustly out-of-distribution. Qualitative analysis reveals that different LLMs employ qualitatively distinct evolutionary strategies: Claude models consistently design high-feature-count linear estimators, while Gpt-oss-120B gravitates toward simpler and more interpretable positional weighting schemes. Surprisingly, only Sonnet 4.5 and Opus 4.5 reliably leverage increased method complexity to improve performance -- Opus 4.6 shows an unexpected regression relative to its predecessor. Overall, our results indicate that LLM-powered evolutionary search is a promising paradigm for automated, interpretable hallucination detector design.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Selective Prediction | TruthfulQA | PRR39.4 | 20 | |
| Selective Prediction | GSM8K | PRR90.4 | 20 | |
| Selective Prediction | bAbI | PRR79.2 | 20 | |
| Selective Prediction | CoQA | PRR80.5 | 20 | |
| Selective Prediction | SamSum | PRR32.8 | 20 | |
| Selective Prediction | WMT fr 14 | Prediction Ranking Rate38.5 | 20 | |
| Selective Prediction | WMT ru 19 | PRR33.3 | 20 | |
| Selective Prediction | MMLU | PRR75.8 | 20 | |
| Selective Prediction | WMT de 14 | Prediction Ranking Rate34.2 | 20 | |
| Selective Prediction | WMT de 19 | PRR45.4 | 20 |