Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

The Geometry of Truth: Emergent Linear Structure in Large Language Model Representations of True/False Datasets

About

Large Language Models (LLMs) have impressive capabilities, but are prone to outputting falsehoods. Recent work has developed techniques for inferring whether a LLM is telling the truth by training probes on the LLM's internal activations. However, this line of work is controversial, with some authors pointing out failures of these probes to generalize in basic ways, among other conceptual issues. In this work, we use high-quality datasets of simple true/false statements to study in detail the structure of LLM representations of truth, drawing on three lines of evidence: 1. Visualizations of LLM true/false statement representations, which reveal clear linear structure. 2. Transfer experiments in which probes trained on one dataset generalize to different datasets. 3. Causal evidence obtained by surgically intervening in a LLM's forward pass, causing it to treat false statements as true and vice versa. Overall, we present evidence that at sufficient scale, LLMs linearly represent the truth or falsehood of factual statements. We also show that simple difference-in-mean probes generalize as well as other probing techniques while identifying directions which are more causally implicated in model outputs.

Samuel Marks, Max Tegmark• 2023

Related benchmarks

TaskDatasetResultRank
Hallucination DetectionTriviaQA
AUROC0.6291
265
Hallucination DetectionTriviaQA (test)
AUC-ROC62.91
169
Hallucination DetectionRAGTruth (test)
AUROC0.6191
83
Hallucination DetectionMATH
Mean AUROC59.58
72
Hallucination DetectionCommonsenseQA
Mean AUROC0.5468
48
Hallucination DetectionCoQA
Mean AUROC0.6161
48
Hallucination DetectionSVAMP
Mean AUROC58.08
48
Hallucination DetectionAverage Cross-domain
Mean AUROC0.5635
48
Hallucination DetectionBelebele
Mean AUROC0.4893
48
Hallucination DetectionRAGTruth
AUROC0.6191
36
Showing 10 of 21 rows

Other info

Follow for update