@inproceedings{tang-etal-2025-top,
title = "Top-$n\sigma$: Eliminating Noise in Logit Space for Robust Token Sampling of {LLM}",
author = "Tang, Chenxia and
Liu, Jianchun and
Xu, Hongli and
Huang, Liusheng",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.528/",
pages = "10758--10774",
ISBN = "979-8-89176-251-0",
abstract = "Large language models (LLMs) rely heavily on sampling methods to generate diverse and high-quality text.While existing sampling methods like top-$p$ and min-$p$ have identified the detrimental effects of low-probability tails in LLMs' outputs, they still fail to effectively distinguish between diversity and noise. This limitation stems from their reliance on probability-based metrics that are inherently sensitive to temperature scaling. Through empirical and theoretical analysis, we make two key discoveries: (1) the pre-softmax logits exhibit a clear statistical separation between informative tokens and noise, and (2) we prove the mathematical equivalence of min-$p$ and top-(1-$p$) under uniform distribution over logits. These findings motivate the design of top-n$\sigma$, a novel sampling method that identifies informative tokens by eliminating noise directly in logit space.Unlike existing methods that become unstable at high temperatures, top-$n\sigma$ achieves temperature-invariant token selection while preserving output diversity. Extensive experiments across reasoning and creative writing tasks demonstrate that our method consistently outperforms existing approaches, with particularly significant improvements in high-temperature settings."
}
Markdown (Informal)
[Top-n𝜎: Eliminating Noise in Logit Space for Robust Token Sampling of LLM](https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.528/) (Tang et al., ACL 2025)
ACL