Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Defending Unauthorized Model Merging via Dual-Stage Weight Protection

About

The rapid proliferation of pretrained models and open repositories has made model merging a convenient yet risky practice, allowing free-riders to combine fine-tuned models into a new multi-capability model without authorization. Such unauthorized model merging not only violates intellectual property rights but also undermines model ownership and accountability. To address this issue, we present MergeGuard, a proactive dual-stage weight protection framework that disrupts merging compatibility while maintaining task fidelity. In the first stage, we redistribute task-relevant information across layers via L2-regularized optimization, ensuring that important gradients are evenly dispersed. In the second stage, we inject structured perturbations to misalign task subspaces, breaking curvature compatibility in the loss landscape. Together, these stages reshape the model's parameter geometry such that merged models collapse into destructive interference while the protected model remains fully functional. Extensive experiments on both vision (ViT-L-14) and language (Llama2, Gemma2, Mistral) models demonstrate that MergeGuard reduces merged model accuracy by up to 90% with less than 1.5% performance loss on the protected model.

Wei-Jia Chen, Min-Yen Tsai, Cheng-Yi Lee, Chia-Mu Yu• 2025

Related benchmarks

TaskDatasetResultRank
Image ClassificationEuroSAT
Accuracy97.4
497
Image ClassificationDTD
Accuracy82.16
419
ClassificationCars
Accuracy90.3
314
Image ClassificationGTSRB
Accuracy98.25
291
Image ClassificationRESISC45
Accuracy97.25
263
Image ClassificationSUN397
Accuracy81.52
246
Image ClassificationMNIST
Accuracy99.27
48
Image ClassificationSVHN
Accuracy96.82
30
Showing 8 of 8 rows

Other info

Follow for update