BLoB: Bayesian Low-Rank Adaptation by Backpropagation for Large Language Models
About
Large Language Models (LLMs) often suffer from overconfidence during inference, particularly when adapted to downstream domain-specific tasks with limited data. Previous work addresses this issue by employing approximate Bayesian estimation after the LLMs are trained, enabling them to quantify uncertainty. However, such post-training approaches' performance is severely limited by the parameters learned during training. In this paper, we go beyond post-training Bayesianization and propose Bayesian Low-Rank Adaptation by Backpropagation (BLoB), an algorithm that continuously and jointly adjusts both the mean and covariance of LLM parameters throughout the whole fine-tuning process. Our empirical results verify the effectiveness of BLoB in terms of generalization and uncertainty estimation, when evaluated on both in-distribution and out-of-distribution data.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Commonsense Reasoning | ARC Challenge | Accuracy68.81 | 190 | |
| Commonsense Reasoning | ARC-C | Accuracy79.42 | 172 | |
| Commonsense Reasoning | OBQA | Accuracy82.73 | 117 | |
| Commonsense Reasoning | ARC-E | Accuracy90.16 | 106 | |
| Paraphrase Detection | MRPC GLUE (val) | Accuracy0.8873 | 27 | |
| Natural Language Inference | RTE (val) | Accuracy0.7605 | 24 | |
| Commonsense Reasoning | WG-S | Accuracy70.89 | 18 | |
| Commonsense Reasoning | BoolQ | Accuracy86.99 | 18 | |
| Commonsense Reasoning | WG-M | Accuracy74.55 | 18 | |
| Common Sense Reasoning | WinoGrande S In-Distribution Llama-3.1-8B (train test) | Accuracy72.36 | 15 |