Anchored Sliding Window: Toward Robust and Imperceptible Linguistic Steganography
About
Linguistic steganography based on language models typically assumes that steganographic texts are transmitted without alteration, making them fragile to even minor modifications. While previous work mitigates this fragility by limiting the context window, it significantly compromises text quality. In this paper, we propose the anchored sliding window (ASW) framework to improve imperceptibility and robustness. In addition to the latest tokens, the prompt and a bridge context are anchored within the context window, encouraging the model to compensate for the excluded tokens. We formulate the optimization of the bridge context as a variant of prompt distillation, which we further extend using self-distillation strategies. Experiments show that our ASW significantly and consistently outperforms the baseline method in text quality, imperceptibility, and robustness across diverse settings. The code is available at github.com/ryehr/ASW_steganography.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Linguistic Steganography | LLM Steganography Evaluation Dataset | Delta PPL0.201 | 24 | |
| Linguistic Steganography | InstructionWild (500 samples) | PPL Delta0.201 | 8 | |
| Linguistic Steganography | InstructionWild | ΔPPL0.201 | 8 | |
| Linguistic Steganography | databricks-dolly-15k | PPL Change (ΔPPL)1.085 | 8 | |
| Linguistic Steganography | Super-NaturalInstructions | ΔPPL5.786 | 8 | |
| Text Fluency Evaluation | Human evaluation (test) | Fluency Score2.56 | 4 |