@inproceedings{zhang-etal-2025-demystify,
title = "Demystify Verbosity Compensation Behavior of Large Language Models",
author = "Zhang, Yusen and
Snigdha Sarathi Das, Sarkar and
Zhang, Rui",
editor = "Noidea, Noidea",
booktitle = "Proceedings of the 2nd Workshop on Uncertainty-Aware NLP (UncertaiNLP 2025)",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.uncertainlp-main.14/",
pages = "160--178",
ISBN = "979-8-89176-349-4",
abstract = "Recent work has revealed Large Language Models (LLMs) often exhibit undesirable behaviors, such as hallucination and toxicity, limiting their reliability and broader adoption. In this paper, we discover an understudied type of undesirable behavior of LLMs, which we term Verbosity Compensation (VC). VC is similar to the hesitation behavior of humans under uncertainty, compensating with excessive words such as repeating questions, introducing ambiguity, or providing excessive enumeration. We present the first work that analyzes Verbosity Compensation, explores its causes, and proposes a simple mitigating approach. Our experiments on five datasets of knowledge and reasoning-based QA tasks with 14 LLMs, reveal three conclusions. 1) A pervasive presence of VC across all models and all datasets. 2) The large performance gap between verbose and concise responses. We also demonstrate that this difference does not naturally diminish as LLM capability increases. 3) Higher uncertainty exhibited by VC responses across all five datasets, suggesting a strong connection between verbosity and model uncertainty. We propose a simple yet effective cascade algorithm that replaces the verbose responses with the other model-generated responses, alleviating the VC of the Mistral model from 63.81{\%} to 16.16{\%} on the Qasper dataset."
}Markdown (Informal)
[Demystify Verbosity Compensation Behavior of Large Language Models](https://preview.aclanthology.org/ingest-emnlp/2025.uncertainlp-main.14/) (Zhang et al., UncertaiNLP 2025)
ACL