Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

MLorc: Momentum Low-rank Compression for Memory Efficient Large Language Model Adaptation

About

With increasing size of large language models (LLMs), full-parameter fine-tuning imposes substantial memory demands. To alleviate this, we propose a novel memory-efficient training paradigm called Momentum Low-rank compression (MLorc). The key idea of MLorc is to compress and reconstruct the momentum of matrix parameters during training to reduce memory consumption. Compared to LoRA, MLorc avoids enforcing a fixed-rank constraint on weight update matrices and thus enables full-parameter learning. Compared to GaLore, MLorc directly compress the momentum rather than gradients, thereby better preserving the training dynamics of full-parameter fine-tuning. We provide a theoretical guarantee for its convergence under mild assumptions. Empirically, MLorc consistently outperforms other memory-efficient training methods, matches or even exceeds the performance of full fine-tuning at small ranks (e.g., $r=4$), and generalizes well across different optimizers, all while not compromising time or memory efficiency.

Wei Shen, Zhang Yaxiang, Minhui Huang, Mengfan Xu, Jiawei Zhang, Cong Shen• 2025

Related benchmarks

TaskDatasetResultRank
Code GenerationHumanEval
Accuracy20.7
99
Natural Language UnderstandingGLUE
CoLA Score62.07
5
Showing 2 of 2 rows

Other info

Follow for update