GradOT: Training-free Gradient-preserving Offsite-tuning for Large Language Models
About
The rapid growth of large language models (LLMs) with traditional centralized fine-tuning emerges as a key technique for adapting these models to domain-specific challenges, yielding privacy risks for both model and data owners. One promising solution, called offsite-tuning (OT), is proposed to address these challenges, where a weaker emulator is compressed from the original model and further fine-tuned with adapter to enhance privacy. However, the existing OT-based methods require high computational costs and lack theoretical analysis. This paper introduces a novel OT approach based on gradient-preserving compression, named GradOT. By analyzing the OT problem through the lens of optimization, we propose a method that selectively applies compression techniques such as rank compression and channel pruning, preserving the gradients of fine-tuned adapters while ensuring privacy. Extensive experiments demonstrate that our approach surpasses existing OT methods, both in terms of privacy protection and model performance. Our method provides a theoretical foundation for OT and offers a practical, training-free solution for offsite-tuning of large-scale LLMs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Question Answering | ARC Challenge | -- | 749 | |
| Question Answering | OpenBookQA | Accuracy30.8 | 465 | |
| Question Answering | ARC Easy | Normalized Acc61.3 | 385 | |
| Physical Interaction Question Answering | PIQA | Accuracy73.6 | 323 | |
| Question Answering | OBQA | Accuracy36.2 | 276 | |
| Question Answering | ARC-C | Accuracy50.1 | 166 | |
| Science Question Answering | ARC-E | Accuracy77.8 | 138 | |
| Sentence Completion | HellaSwag | Accuracy41.7 | 133 | |
| Multiple-choice Question Answering | SciQ | Accuracy93.9 | 74 | |
| Question Answering | WebQuestions (WebQs) | Accuracy48.1 | 67 |