Sparse Autoencoders for Hypothesis Generation
About
We describe HypotheSAEs, a general method to hypothesize interpretable relationships between text data (e.g., headlines) and a target variable (e.g., clicks). HypotheSAEs has three steps: (1) train a sparse autoencoder on text embeddings to produce interpretable features describing the data distribution, (2) select features that predict the target variable, and (3) generate a natural language interpretation of each feature (e.g., "mentions being surprised or shocked") using an LLM. Each interpretation serves as a hypothesis about what predicts the target variable. Compared to baselines, our method better identifies reference hypotheses on synthetic datasets (at least +0.06 in F1) and produces more predictive hypotheses on real datasets (~twice as many significant findings), despite requiring 1-2 orders of magnitude less compute than recent LLM-based methods. HypotheSAEs also produces novel discoveries on two well-studied tasks: explaining partisan differences in Congressional speeches and identifying drivers of engagement with online headlines.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Signal Recovery | Synthetic dataset | SURF Score90 | 5 | |
| Hypothesis Discovery | HypoBench | Deception Score8 | 3 | |
| Hypothesis Discovery | Significant Hypotheses Count12 | 3 | ||
| Hypothesis Discovery | Design | Significant Hypotheses Found6 | 3 | |
| Hypothesis Discovery | Congress | Significant Hypotheses Found12 | 3 | |
| Hypothesis Discovery | CMV | Significant Hypotheses Found8 | 3 | |
| Hypothesis Discovery | LaMem | Significant Hypotheses0.00e+0 | 3 |