Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

DiaBlo: Diagonal Blocks Are Sufficient For Finetuning

About

Fine-tuning is a critical step for adapting large language models (LLMs) to domain-specific downstream tasks. To mitigate the substantial computational and memory costs of full-model fine-tuning, Parameter-Efficient Fine-Tuning (PEFT) methods have been proposed to update only a small subset of model parameters. However, performance gaps between PEFT approaches and full-model fine-tuning still exist. In this work, we present DiaBlo, a simple yet effective PEFT approach that updates only the diagonal blocks of selected model weight matrices. Unlike Low-Rank Adaptation (LoRA) and its variants, DiaBlo eliminates the need for low-rank matrix products, thereby avoiding the reliance on auxiliary initialization schemes or customized optimization strategies to improve convergence. This design leads to stable and robust convergence while maintaining comparable memory efficiency and training speed to LoRA. Moreover, we provide theoretical guarantees showing that, under mild low-rank conditions, DiaBlo is more expressive than LoRA in the linear problem and converges to a stationary point of the general nonlinear full fine-tuning. Through extensive experiments across a range of tasks, including commonsense reasoning, arithmetic reasoning, code generation, and safety alignment, we show that fine-tuning only diagonal blocks is sufficient for strong and consistent performance. DiaBlo not only achieves competitive accuracy but also preserves high memory efficiency and fast fine-tuning speed. Codes are available at https://github.com/ziyangjoy/DiaBlo.

Selcuk Gurses, Aozhong Zhang, Yanxia Deng, Xun Dong, Xin Li, Naigang Wang, Penghang Yin, Zi Yang• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval (test)
Pass@143.2
506
Commonsense ReasoningCommon Sense Reasoning Tasks
Avg Score88.3
316
Arithmetic ReasoningGSM8K
Accuracy66.5
173
Commonsense ReasoningCommonsense Reasoning (BoolQ, PIQA, SIQA, HellaS., WinoG., ARC-e, ARC-c, OBQA)
BoolQ Accuracy76.1
129
Arithmetic ReasoningAQuA, GSM8K, MAWPS, SVAMP
AQuA Accuracy27.6
31
Arithmetic ReasoningMATH
Accuracy20.4
23
Safety AlignmentHEX-PHI
HEx-PHI Score98.8
12
Arithmetic ReasoningGSM8K and MATH Average
Average Accuracy43.4
7
Multi-turn Dialogue EvaluationMT-Bench (test)
MT-Bench Score6.26
6
Natural Language UnderstandingGLUE (test)
MRPC Score86
5
Showing 10 of 10 rows

Other info

Follow for update