Adaptive Activation Steering: A Tuning-Free LLM Truthfulness Improvement Method for Diverse Hallucinations Categories
About
Recent studies have indicated that Large Language Models (LLMs) harbor an inherent understanding of truthfulness, yet often fail to consistently express it and generate false statements. This gap between "knowing" and "telling" poses a challenge for ensuring the truthfulness of generated content. Inspired by recent work on the practice of encoding human-interpretable concepts linearly within large language models, we treat truthfulness as a specially linearly encoded concept within LLMs, and introduce Adaptive Activation Steering (ACT), a tuning-free method that adaptively shifts LLM's activations in the "truthful" direction during inference. ACT addresses diverse categories of hallucinations by utilizing diverse truthfulness-related steering vectors and adjusting the steering intensity adaptively. Applied as an add-on across various models, ACT significantly improves truthfulness in LLaMA ($\uparrow$ 142%), LLaMA2 ($\uparrow$ 24%), Alpaca ($\uparrow$ 36%), Vicuna ($\uparrow$ 28%), LLaMA2-Chat ($\uparrow$ 19%), and LLaMA3($\uparrow$ 34%). Furthermore, we verify ACT's scalability across larger models (13B, 33B, 65B), underscoring the adaptability of ACT to large-scale language models. Our code is available at https://github.com/tianlwang/ACT.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Commonsense Reasoning | HellaSwag | Accuracy70.1 | 1891 | |
| Mathematical Reasoning | GSM8K | Accuracy87.9 | 1362 | |
| Massive Multitask Language Understanding | MMLU | Accuracy59.4 | 117 | |
| Factuality Evaluation | TruthfulQA | MC293.7 | 73 | |
| Safety Refusal | AdvBench | Refusal Rate83.9 | 46 | |
| AI Text Detection | HC3 AI-Prob | Finance Accuracy56.3 | 24 | |
| Sentiment Analysis | SST-2 | Positive Rate48.6 | 24 | |
| Safety Refusal | Jailbreak Prompts | Refusal Rate78.5 | 15 | |
| Safety Refusal | ToxicChat | Refusal Rate59.8 | 15 |