Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Instructional Fingerprinting of Large Language Models

About

The exorbitant cost of training Large language models (LLMs) from scratch makes it essential to fingerprint the models to protect intellectual property via ownership authentication and to ensure downstream users and developers comply with their license terms (e.g. restricting commercial use). In this study, we present a pilot study on LLM fingerprinting as a form of very lightweight instruction tuning. Model publisher specifies a confidential private key and implants it as an instruction backdoor that causes the LLM to generate specific text when the key is present. Results on 11 popularly-used LLMs showed that this approach is lightweight and does not affect the normal behavior of the model. It also prevents publisher overclaim, maintains robustness against fingerprint guessing and parameter-efficient training, and supports multi-stage fingerprinting akin to MIT License. Code is available in https://cnut1648.github.io/Model-Fingerprint/.

Jiashu Xu, Fei Wang, Mingyu Derek Ma, Pang Wei Koh, Chaowei Xiao, Muhao Chen• 2024

Related benchmarks

TaskDatasetResultRank
Mathematical ReasoningGSM8K
Math Score42
171
Mathematical ReasoningMGSM
Accuracy42
114
Safety EvaluationToxigen
Safety50
71
Fingerprint RemovalLLM Fingerprinting Evaluation Alpaca-GPT4-52k
ASR Error Rate0.00e+0
66
Fingerprint VerificationShisa-7B and Abel-7B-002 Merged
VSR1
60
Japanese Language UnderstandingJAQKET
Japanese Score78
60
Mathematical ReasoningWizardMath (test)
Math Score43
60
Fingerprint VerificationFingerprint Verification
VSR100
60
Fingerprint VerificationEmbedded Fingerprints (test)
VSR1
60
Safety EvaluationLLaMA-2-7B-CHAT Safety (test)
Safety Score0.5
60
Showing 10 of 38 rows

Other info

Follow for update