QuantileMark: A Message-Symmetric Multi-bit Watermark for LLMs
About
As large language models become standard backends for content generation, practical provenance increasingly requires multi-bit watermarking. In provider-internal deployments, a key requirement is message symmetry: the message itself should not systematically affect either text quality or verification outcomes. Vocabulary-partition watermarks can break message symmetry in low-entropy decoding: some messages are assigned most of the probability mass, while others are forced to use tail tokens. This makes embedding quality and message decoding accuracy message-dependent. We propose QuantileMark, a white-box multi-bit watermark that embeds messages within the continuous cumulative probability interval $[0, 1)$. At each step, QuantileMark partitions this interval into $M$ equal-mass bins and samples strictly from the bin assigned to the target symbol, ensuring a fixed $1/M$ probability budget regardless of context entropy. For detection, the verifier reconstructs the same partition under teacher forcing, computes posteriors over latent bins, and aggregates evidence for verification. We prove message-unbiasedness, a property ensuring that the base distribution is recovered when averaging over messages. This provides a theoretical foundation for generation-side symmetry, while the equal-mass design additionally promotes uniform evidence strength across messages on the detection side. Empirical results on C4 continuation and LFQA show improved multi-bit recovery and detection robustness over strong baselines, with negligible impact on generation quality. Our code is available at GitHub (https://github.com/zzzjunlin/QuantileMark).
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-bit Watermarking | C4 | Perplexity7.404 | 4 | |
| Summarization | CNN/DailyMail | R-1 Score37.44 | 4 | |
| Answer quality evaluation | LFQA | GPT-4o Score4.109 | 4 | |
| Machine Translation | WMT'14 | BLEU17.41 | 4 | |
| Multi-bit Watermarking | LFQA | Perplexity2.759 | 4 | |
| Generative Watermarking | C4 No attack | Bit Accuracy98.93 | 3 | |
| Generative Watermarking | C4 Copy-paste, epsilon = 0.2 | Bit Accuracy97.3 | 3 | |
| Generative Watermarking | C4 Synonym, epsilon = 0.2 | Bit Accuracy97.12 | 3 | |
| Generative Watermarking | C4 Paraphrase Dipper | Bit Accuracy76.4 | 3 | |
| Generative Watermarking | C4 Deletion, epsilon = 0.1 | Bit Accuracy87.12 | 3 |