Bayesian Low-rank Adaptation for Large Language Models
About
Low-rank adaptation (LoRA) has emerged as a new paradigm for cost-efficient fine-tuning of large language models (LLMs). However, fine-tuned LLMs often become overconfident especially when fine-tuned on small datasets. Bayesian methods, with their inherent ability to estimate uncertainty, serve as potent tools to mitigate overconfidence and enhance calibration. In this work, we introduce Laplace-LoRA, which applies a Bayesian approach to the LoRA parameters. Specifically, Laplace-LoRA applies a Laplace approximation to the posterior over the LoRA parameters, considerably improving the calibration of fine-tuned LLMs.
Adam X. Yang, Maxime Robeyns, Xi Wang, Laurence Aitchison• 2023
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Classification | DTD | -- | 487 | |
| Image Classification | Flowers102 | Accuracy91.5 | 478 | |
| Commonsense Reasoning | ARC Challenge | Accuracy66.9 | 132 | |
| Commonsense Reasoning | OBQA | Accuracy82.12 | 75 | |
| Commonsense Reasoning | ARC-E | Accuracy85.4 | 62 | |
| Commonsense Reasoning | ARC-C | Accuracy66.78 | 51 | |
| Paraphrase Detection | MRPC GLUE (val) | Accuracy0.8652 | 27 | |
| Natural Language Inference | RTE (val) | Accuracy0.7605 | 24 | |
| Commonsense Reasoning | WG-M | Accuracy75.55 | 18 | |
| Commonsense Reasoning | WG-S | Accuracy69.2 | 18 |
Showing 10 of 31 rows