Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Bayesian Low-rank Adaptation for Large Language Models

About

Low-rank adaptation (LoRA) has emerged as a new paradigm for cost-efficient fine-tuning of large language models (LLMs). However, fine-tuned LLMs often become overconfident especially when fine-tuned on small datasets. Bayesian methods, with their inherent ability to estimate uncertainty, serve as potent tools to mitigate overconfidence and enhance calibration. In this work, we introduce Laplace-LoRA, which applies a Bayesian approach to the LoRA parameters. Specifically, Laplace-LoRA applies a Laplace approximation to the posterior over the LoRA parameters, considerably improving the calibration of fine-tuned LLMs.

Adam X. Yang, Maxime Robeyns, Xi Wang, Laurence Aitchison• 2023

Related benchmarks

TaskDatasetResultRank
Image ClassificationFlowers102
Accuracy91.5
558
Image ClassificationDTD--
542
Commonsense ReasoningARC Challenge
Accuracy66.9
190
Commonsense ReasoningARC-C
Accuracy81.05
172
Commonsense ReasoningOBQA
Accuracy82.12
117
Commonsense ReasoningARC-E
Accuracy90.66
106
Multiple-choice Question AnsweringOBQA
Accuracy92.53
69
Multiple-choice Question AnsweringRACE
Accuracy88.51
54
Out-of-Distribution DetectionRACE to MMLU
AUROC53.43
41
Out-of-Distribution DetectionOBQA to MMLU
AUROC51.54
41
Showing 10 of 53 rows

Other info

Follow for update