UNCERTAINTY-LINE: Length-Invariant Estimation of Uncertainty for Large Language Models

Roman Vashurin, Maiya Goloburda, Preslav Nakov, Maxim Panov


Abstract
Large Language Models (LLMs) have become indispensable tools across various applications, making it more important than ever to ensure the quality and the trustworthiness of their outputs. This has led to growing interest in uncertainty quantification (UQ) methods for assessing the reliability of LLM outputs. Many existing UQ techniques rely on token probabilities, which inadvertently introduces a bias with respect to the length of the output. While some methods attempt to account for this, we demonstrate that such biases persist even in length-normalized approaches. To address the problem, here we propose UNCERTAINTY-LINE (Length-INvariant Estimation), a simple debiasing procedure that regresses uncertainty scores on output length and uses the residuals as corrected, length-invariant estimates. Our method is post-hoc, model-agnostic, and applicable to a range of UQ measures. Through extensive evaluation on machine translation, summarization, and question-answering tasks, we demonstrate that UNCERTAINTY-LINE consistently improves over even nominally length-normalized UQ methods uncertainty estimates across multiple metrics and models. We release our code publicly at https://github.com/stat-ml/uncertainty-line
Anthology ID:
2025.emnlp-main.400
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7892–7919
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.400/
DOI:
Bibkey:
Cite (ACL):
Roman Vashurin, Maiya Goloburda, Preslav Nakov, and Maxim Panov. 2025. UNCERTAINTY-LINE: Length-Invariant Estimation of Uncertainty for Large Language Models. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 7892–7919, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
UNCERTAINTY-LINE: Length-Invariant Estimation of Uncertainty for Large Language Models (Vashurin et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.400.pdf
Checklist:
 2025.emnlp-main.400.checklist.pdf