@inproceedings{sicilia-alikhani-2025-adaptive,
title = "Adaptive {P}latt Scaling with Causal Interpretations for Self-Reflective Language Model Uncertainty Estimates",
author = "Sicilia, Anthony and
Alikhani, Malihe",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.999/",
doi = "10.18653/v1/2025.findings-emnlp.999",
pages = "18414--18422",
ISBN = "979-8-89176-335-7",
abstract = "As large language models (LLMs) are consumed by more users and deployed in increasingly autonomous capacities, their ability to self-monitor and ask for human intervention is of vital importance. Underlying this capability are fundamental skills like self-reflection and expression of uncertainty. In this work, we provide a formal analysis of LLM self-reflection for uncertainty estimation, using domain adaptation theory to model the shift between base predictions and reflective judgments. We use this to motivate a temperature scaling algorithm that calibrates uncertainty using comparisons between base predictions and LLM self-reflections. We evaluate our approach on challenging question-answering tasks requiring reasoning, demonstrating that our methods can improve calibration of uncertainty estimates and also offer improvements in human interpretation. More broadly, this use case shows how domain adaptation presents a promising analytical tool for understanding the underlying statistical properties of LLM self-reflections."
}Markdown (Informal)
[Adaptive Platt Scaling with Causal Interpretations for Self-Reflective Language Model Uncertainty Estimates](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.999/) (Sicilia & Alikhani, Findings 2025)
ACL