Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

GradOT: Training-free Gradient-preserving Offsite-tuning for Large Language Models

About

The rapid growth of large language models (LLMs) with traditional centralized fine-tuning emerges as a key technique for adapting these models to domain-specific challenges, yielding privacy risks for both model and data owners. One promising solution, called offsite-tuning (OT), is proposed to address these challenges, where a weaker emulator is compressed from the original model and further fine-tuned with adapter to enhance privacy. However, the existing OT-based methods require high computational costs and lack theoretical analysis. This paper introduces a novel OT approach based on gradient-preserving compression, named GradOT. By analyzing the OT problem through the lens of optimization, we propose a method that selectively applies compression techniques such as rank compression and channel pruning, preserving the gradients of fine-tuned adapters while ensuring privacy. Extensive experiments demonstrate that our approach surpasses existing OT methods, both in terms of privacy protection and model performance. Our method provides a theoretical foundation for OT and offers a practical, training-free solution for offsite-tuning of large-scale LLMs.

Kai Yao, Zhaorui Tan, Penglei Gao, Lichun Li, Kaixin Wu, Yinggui Wang, Yuan Zhao, Yixin Ji, Wei Wang, Jianke Zhu• 2025

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC Challenge--
749
Question AnsweringOpenBookQA
Accuracy30.8
465
Question AnsweringARC Easy
Normalized Acc61.3
385
Physical Interaction Question AnsweringPIQA
Accuracy73.6
323
Question AnsweringOBQA
Accuracy36.2
276
Question AnsweringARC-C
Accuracy50.1
166
Science Question AnsweringARC-E
Accuracy77.8
138
Sentence CompletionHellaSwag
Accuracy41.7
133
Multiple-choice Question AnsweringSciQ
Accuracy93.9
74
Question AnsweringWebQuestions (WebQs)
Accuracy48.1
67
Showing 10 of 12 rows

Other info

Code

Follow for update