Unveiling Confirmation Bias in Chain-of-Thought Reasoning

Yue Wan, Xiaowei Jia, Xiang Lorraine Li


Abstract
Chain-of-thought (CoT) prompting has been widely adopted to enhance the reasoning capabilities of large language models (LLMs). However, the effectiveness of CoT reasoning is inconsistent across tasks with different reasoning types. This work presents a novel perspective to understand CoT behavior through the lens of confirmation bias in cognitive psychology. Specifically, we examine how model internal beliefs, approximated by direct question-answering probabilities, affect both reasoning generation (Q → R) and reasoning-guided answer prediction (QR → A) in CoT. By decomposing CoT into a two-stage process, we conduct a thorough correlation analysis in model beliefs, rationale attributes, and stage-wise performance. Our results provide strong evidence of confirmation bias in LLMs, such that model beliefs not only skew the reasoning process but also influence how rationales are utilized for answer prediction. Furthermore, the interplay between task vulnerability to confirmation bias and the strength of beliefs also provides explanations for CoT effectiveness across reasoning tasks and models. Overall, this study provides a valuable insight for the needs of better prompting strategies that mitigate confirmation bias to enhance reasoning performance. Code is available at https://github.com/yuewan2/biasedcot.
Anthology ID:
2025.findings-acl.195
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3788–3804
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.195/
DOI:
Bibkey:
Cite (ACL):
Yue Wan, Xiaowei Jia, and Xiang Lorraine Li. 2025. Unveiling Confirmation Bias in Chain-of-Thought Reasoning. In Findings of the Association for Computational Linguistics: ACL 2025, pages 3788–3804, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Unveiling Confirmation Bias in Chain-of-Thought Reasoning (Wan et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.195.pdf