All Roads Lead to Rome: Graph-Based Confidence Estimation for Large Language Model Reasoning

Caiqi Zhang, Chang Shu, Ehsan Shareghi, Nigel Collier


Abstract
Confidence estimation is essential for the reliable deployment of large language models (LLMs). Existing methods are primarily designed for factual QA tasks and often fail to generalize to reasoning tasks. To address this gap, we propose a set of training-free, graph-based confidence estimation methods tailored to reasoning tasks. Our approach models reasoning paths as directed graphs and estimates confidence by exploiting graph properties such as centrality, path convergence, and path weighting. Experiments with two LLMs on three reasoning datasets demonstrate improved confidence estimation and enhanced performance on two downstream tasks.
Anthology ID:
2025.emnlp-main.1620
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
31802–31812
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1620/
DOI:
Bibkey:
Cite (ACL):
Caiqi Zhang, Chang Shu, Ehsan Shareghi, and Nigel Collier. 2025. All Roads Lead to Rome: Graph-Based Confidence Estimation for Large Language Model Reasoning. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 31802–31812, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
All Roads Lead to Rome: Graph-Based Confidence Estimation for Large Language Model Reasoning (Zhang et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1620.pdf
Checklist:
 2025.emnlp-main.1620.checklist.pdf