@inproceedings{huang-etal-2025-enhancing-llm,
title = "Enhancing {LLM} Text Detection with Retrieved Contexts and Logits Distribution Consistency",
author = "Huang, Zhaoheng and
Zhu, Yutao and
Wen, Ji-Rong and
Dou, Zhicheng",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.503/",
pages = "9933--9945",
ISBN = "979-8-89176-332-6",
abstract = "Large language models (LLMs) can generate fluent text, raising concerns about misuse in online comments and academic writing, leading to issues like corpus pollution and copyright infringement. Existing LLM text detection methods often rely on features from the logit distribution of the input text. However, the distinction between the LLM-generated and human-written texts may rely on only a few tokens due to the short length or insufficient information in some texts, leading to minimal and hard-to-detect differences in logit distributions. To address this, we propose HALO, an LLM-based detection method that leverages external text corpora to evaluate the difference in the logit distribution of input text under retrieved human-written and LLM-rewritten contexts. HALO also complements basic detection features and can serve as a plug-and-play module to enhance existing detection methods. Extensive experiments on five public datasets with three widely-used source LLMs show that our proposed detection method achieves state-of-the-art performance in AUROC, both in cross-domain and domain-specific scenarios."
}Markdown (Informal)
[Enhancing LLM Text Detection with Retrieved Contexts and Logits Distribution Consistency](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.503/) (Huang et al., EMNLP 2025)
ACL