Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

ScaleOT: Privacy-utility-scalable Offsite-tuning with Dynamic LayerReplace and Selective Rank Compression

About

Offsite-tuning is a privacy-preserving method for tuning large language models (LLMs) by sharing a lossy compressed emulator from the LLM owners with data owners for downstream task tuning. This approach protects the privacy of both the model and data owners. However, current offsite tuning methods often suffer from adaptation degradation, high computational costs, and limited protection strength due to uniformly dropping LLM layers or relying on expensive knowledge distillation. To address these issues, we propose ScaleOT, a novel privacy-utility-scalable offsite-tuning framework that effectively balances privacy and utility. ScaleOT introduces a novel layerwise lossy compression algorithm that uses reinforcement learning to obtain the importance of each layer. It employs lightweight networks, termed harmonizers, to replace the raw LLM layers. By combining important original LLM layers and harmonizers in different ratios, ScaleOT generates emulators tailored for optimal performance with various model scales for enhanced privacy protection. Additionally, we present a rank reduction method to further compress the original LLM layers, significantly enhancing privacy with negligible impact on utility. Comprehensive experiments show that ScaleOT can achieve nearly lossless offsite tuning performance compared with full fine-tuning while obtaining better model privacy.

Kai Yao, Zhaorui Tan, Tiandi Ye, Lichun Li, Yuan Zhao, Wenyan Liu, Wei Wang, Jianke Zhu• 2024

Related benchmarks

TaskDatasetResultRank
Question AnsweringARC Challenge--
749
Question AnsweringOpenBookQA
Accuracy28.2
465
Question AnsweringARC Easy
Normalized Acc61.9
385
Physical Interaction Question AnsweringPIQA
Accuracy75.2
323
Sentence CompletionHellaSwag
Accuracy42.9
133
Multiple-choice Question AnsweringSciQ
Accuracy94
74
Question AnsweringWebQuestions (WebQs)
Accuracy28.2
67
Multiple-choice Question AnsweringRACE
Accuracy40.8
46
Question Answering8 Question Answering Benchmarks (OBQA, PIQA, ARC-E, ARC-C, HellaSwag, SciQ, WebQs, RACE) (test)
OBQA37.4
14
Showing 9 of 9 rows

Other info

Follow for update