@inproceedings{silva-etal-2024-look,
title = "Look Who`s Talking Now: Covert Channels From Biased {LLM}s",
author = "Silva, Daniel and
Sala, Frederic and
Gabrys, Ryan",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.971/",
doi = "10.18653/v1/2024.findings-emnlp.971",
pages = "16648--16658",
abstract = "Large language model-based steganography encodes hidden messages into model-generated tokens. The key tradeoff is between how much hidden information can be introduced and how much the model can be perturbed. To address this tradeoff, we show how to adapt strategies previously used for LLM watermarking to encode large amounts of information. We tackle the practical (but difficult) setting where we do not have access to the full model when trying to recover the hidden information. Theoretically, we study the fundamental limits in how much steganographic information can be inserted into LLM-created outputs. We provide practical encoding schemes and present experimental results showing that our proposed strategies are nearly optimal."
}
Markdown (Informal)
[Look Who’s Talking Now: Covert Channels From Biased LLMs](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.971/) (Silva et al., Findings 2024)
ACL