@inproceedings{dubois-etal-2025-sampling,
title = "How Sampling Affects the Detectability of Machine-written texts: A Comprehensive Study",
author = "Dubois, Matthieu and
Yvon, Fran{\c{c}}ois and
Piantanida, Pablo",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.609/",
doi = "10.18653/v1/2025.findings-emnlp.609",
pages = "11369--11387",
ISBN = "979-8-89176-335-7",
abstract = "As texts generated by Large Language Models (LLMs) are ever more common and often indistinguishable from human-written content, research on automatic text detection has attracted growing attention. Many recent detectors report near-perfect accuracy, often boasting AUROC scores above 99{\%}. However, these claims typically assume fixed generation settings, leaving open the question of how robust such systems are to changes in decoding strategies. In this work, we systematically examine how sampling-based decoding impacts detectability, with a focus on how subtle variations in a model{'}s (sub)word-level distribution affect detection performance. We find that even minor adjustments to decoding parameters - such as temperature, top-p, or nucleus sampling - can severely impair detector accuracy, with AUROC dropping from near-perfect levels to 1{\%} in some settings. Our findings expose critical blind spots in current detection methods and emphasize the need for more comprehensive evaluation protocols. To facilitate future research, we release a large-scale dataset encompassing 37 decoding configurations, along with our code and evaluation framework \url{https://github.com/BaggerOfWords/Sampling-and-Detection}."
}Markdown (Informal)
[How Sampling Affects the Detectability of Machine-written texts: A Comprehensive Study](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.609/) (Dubois et al., Findings 2025)
ACL