Bayesian Low-rank Adaptation for Large Language Models
About
Low-rank adaptation (LoRA) has emerged as a new paradigm for cost-efficient fine-tuning of large language models (LLMs). However, fine-tuned LLMs often become overconfident especially when fine-tuned on small datasets. Bayesian methods, with their inherent ability to estimate uncertainty, serve as potent tools to mitigate overconfidence and enhance calibration. In this work, we introduce Laplace-LoRA, which applies a Bayesian approach to the LoRA parameters. Specifically, Laplace-LoRA applies a Laplace approximation to the posterior over the LoRA parameters, considerably improving the calibration of fine-tuned LLMs.
Adam X. Yang, Maxime Robeyns, Xi Wang, Laurence Aitchison• 2023
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | Flowers102 | Accuracy91.5 | 558 | |
| Image Classification | DTD | -- | 542 | |
| Commonsense Reasoning | ARC Challenge | Accuracy66.9 | 190 | |
| Commonsense Reasoning | ARC-C | Accuracy81.05 | 172 | |
| Commonsense Reasoning | OBQA | Accuracy82.12 | 117 | |
| Commonsense Reasoning | ARC-E | Accuracy90.66 | 106 | |
| Multiple-choice Question Answering | OBQA | Accuracy92.53 | 69 | |
| Multiple-choice Question Answering | RACE | Accuracy88.51 | 54 | |
| Out-of-Distribution Detection | RACE to MMLU | AUROC53.43 | 41 | |
| Out-of-Distribution Detection | OBQA to MMLU | AUROC51.54 | 41 |
Showing 10 of 53 rows