Token-Specific Watermarking with Enhanced Detectability and Semantic Coherence for Large Language Models
About
Large language models generate high-quality responses with potential misinformation, underscoring the need for regulation by distinguishing AI-generated and human-written texts. Watermarking is pivotal in this context, which involves embedding hidden markers in texts during the LLM inference phase, which is imperceptible to humans. Achieving both the detectability of inserted watermarks and the semantic quality of generated texts is challenging. While current watermarking algorithms have made promising progress in this direction, there remains significant scope for improvement. To address these challenges, we introduce a novel multi-objective optimization (MOO) approach for watermarking that utilizes lightweight networks to generate token-specific watermarking logits and splitting ratios. By leveraging MOO to optimize for both detection and semantic objective functions, our method simultaneously achieves detectability and semantic integrity. Experimental results show that our method outperforms current watermarking techniques in enhancing the detectability of texts generated by LLMs while maintaining their semantic coherence. Our code is available at https://github.com/mignonjia/TS_watermark.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Watermark Detection | Llama-3 8B Instruct 30 tokens (generations) | Mean Precision9 | 13 | |
| Watermark Detection | Llama-3-8B-Instruct 150 tokens (generations) | Mean P0.35 | 13 | |
| Watermark Detection | Llama-3-8B Swap perturbation, 30 tokens 1.0 (test) | Mean P0.0092 | 6 | |
| Watermark Detection | Llama-3-8B Swap perturbation, 150 tokens 1.0 (test) | Mean P4.20e-6 | 6 | |
| Watermark Detection Robustness | Llama-3-8B Swap 30%, 150 Tokens | Mean P3.70e-4 | 6 | |
| Watermark Detection Robustness | Llama-3-8B Swap 50%, 150 Tokens | Mean P0.032 | 6 | |
| Watermark Detection | Llama-3-8B Delete perturbation, 30 tokens 1.0 (test) | Mean P0.0043 | 6 | |
| Watermark Detection | Llama-3-8B Delete perturbation, 150 tokens 1.0 (test) | Mean P1.50e-7 | 6 | |
| Watermark Detection | Llama-3-8B Translate perturbation, 30 tokens 1.0 (test) | Mean P0.022 | 6 | |
| Watermark Detection | Llama-3-8B Translate perturbation, 150 tokens 1.0 (test) | Mean P4.20e-5 | 6 |