Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

3-in-1: 2D Rotary Adaptation for Efficient Finetuning, Efficient Batching and Composability

About

Parameter-efficient finetuning (PEFT) methods effectively adapt large language models (LLMs) to diverse downstream tasks, reducing storage and GPU memory demands. Despite these advantages, several applications pose new challenges to PEFT beyond mere parameter efficiency. One notable challenge involves the efficient deployment of LLMs equipped with multiple task- or user-specific adapters, particularly when different adapters are needed for distinct requests within the same batch. Another challenge is the interpretability of LLMs, which is crucial for understanding how LLMs function. Previous studies introduced various approaches to address different challenges. In this paper, we introduce a novel method, RoAd, which employs a straightforward 2D rotation to adapt LLMs and addresses all the above challenges: (1) RoAd is remarkably parameter-efficient, delivering optimal performance on GLUE, eight commonsense reasoning tasks and four arithmetic reasoning tasks with $<0.1\%$ trainable parameters; (2) RoAd facilitates the efficient serving of requests requiring different adapters within a batch, with an overhead comparable to element-wise multiplication instead of batch matrix multiplication; (3) RoAd enhances LLM's interpretability through integration within a framework of distributed interchange intervention, demonstrated via composition experiments.

Baohao Liao, Christof Monz• 2024

Related benchmarks

TaskDatasetResultRank
Natural Language UnderstandingGLUE (dev)
SST-2 (Acc)96.3
518
Instruction FollowingAlpacaEval 2.0
Win Rate62.64
507
Image ClassificationVTAB 1K
Overall Mean Accuracy57.5
258
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA) (test)
BoolQ Accuracy73.3
202
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA)
BoolQ Accuracy74.6
129
Arithmetic ReasoningGSM8K (test)
Accuracy40.7
129
Mathematical ReasoningMAWPS (test)
Accuracy84.9
87
Arithmetic ReasoningAQuA (test)
Accuracy26.8
58
Arithmetic ReasoningSVAMP (test)
Accuracy59.5
54
Visual Instruction TuningLLaVA Evaluation Suite (GQA, SQA, VQAT, POPE) 1.5 (test)
GQA62.5
3
Showing 10 of 10 rows

Other info

Code

Follow for update