Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Bayesian Low-rank Adaptation for Large Language Models

About

Low-rank adaptation (LoRA) has emerged as a new paradigm for cost-efficient fine-tuning of large language models (LLMs). However, fine-tuned LLMs often become overconfident especially when fine-tuned on small datasets. Bayesian methods, with their inherent ability to estimate uncertainty, serve as potent tools to mitigate overconfidence and enhance calibration. In this work, we introduce Laplace-LoRA, which applies a Bayesian approach to the LoRA parameters. Specifically, Laplace-LoRA applies a Laplace approximation to the posterior over the LoRA parameters, considerably improving the calibration of fine-tuned LLMs.

Adam X. Yang, Maxime Robeyns, Xi Wang, Laurence Aitchison• 2023

Related benchmarks

TaskDatasetResultRank
Image ClassificationDTD--
487
Image ClassificationFlowers102
Accuracy91.5
478
Commonsense ReasoningARC Challenge
Accuracy66.9
132
Commonsense ReasoningOBQA
Accuracy82.12
75
Commonsense ReasoningARC-E
Accuracy85.4
62
Commonsense ReasoningARC-C
Accuracy66.78
51
Paraphrase DetectionMRPC GLUE (val)
Accuracy0.8652
27
Natural Language InferenceRTE (val)
Accuracy0.7605
24
Commonsense ReasoningWG-M
Accuracy75.55
18
Commonsense ReasoningWG-S
Accuracy69.2
18
Showing 10 of 31 rows

Other info

Follow for update