Counterfactual-Consistency Prompting for Relative Temporal Understanding in Large Language Models

Jongho Kim, Seung-won Hwang


Abstract
Despite the advanced capabilities of large language models (LLMs), their temporal reasoning ability remains underdeveloped. Prior works have highlighted this limitation, particularly in maintaining temporal consistency when understanding event relations. For example, models often confuse mutually exclusive temporal relations like “before” and “after” between events and make inconsistent predictions. In this work, we tackle the issue of temporal inconsistency in LLMs by proposing a novel counterfactual prompting approach. Our method generates counterfactual questions and enforces collective constraints, enhancing the model’s consistency. We evaluate our method on multiple datasets, demonstrating significant improvements in event ordering for explicit and implicit events and temporal commonsense understanding, by effectively addressing temporal inconsistencies.
Anthology ID:
2025.acl-short.97
Volume:
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1210–1225
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.acl-short.97/
DOI:
Bibkey:
Cite (ACL):
Jongho Kim and Seung-won Hwang. 2025. Counterfactual-Consistency Prompting for Relative Temporal Understanding in Large Language Models. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 1210–1225, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Counterfactual-Consistency Prompting for Relative Temporal Understanding in Large Language Models (Kim & Hwang, ACL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.acl-short.97.pdf