@inproceedings{abbood-etal-2025-time,
title = "Time to Revisit Exact Match",
author = "Abbood, Auss and
Meng, Zaiqiao and
Collier, Nigel",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.637/",
doi = "10.18653/v1/2025.findings-emnlp.637",
pages = "11903--11926",
ISBN = "979-8-89176-335-7",
abstract = "Temporal question answering is an established method for evaluating temporal reasoning in large language models. Expected answers are often numeric (e.g., dates or durations), yet model responses are evaluated like regular text with exact match (EM), unable to distinguish small from large errors. In this investigative work, we frame temporal question answering as a numerical estimation task to assess the shortcomings of EM. We introduce \textit{TempAnswerQA}, a benchmark distilled from Test of Time and TempTabQA, where all questions require a numerical, temporal answer, allowing us to evaluate models beyond EM. We use the forecasting metrics symmetric mean absolute percentage error (sMAPE) and mean absolute scaled error (MASE). With sMAPE, we find that error size and EM are decoupled. Models with low EM still have low sMAPE (both 20{\%}), and some models have high sMAPE despite high EM. Scaling errors by the deviation of the ground truth data with MASE reshuffles model rankings compared to EM, revealing gaps in models' understanding of temporal domain knowledge, especially when trained with synthetic data. Lastly, the models' most frequent error is to deviate by only $\pm1$ from the ground truth. sMAPE and MASE, unlike EM, adequately weight these errors. Our findings underscore the need for specialised metrics for temporal QA tasks. Our code and data are available on https://github.com/aauss/temporal-answer-qa."
}Markdown (Informal)
[Time to Revisit Exact Match](https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.637/) (Abbood et al., Findings 2025)
ACL
- Auss Abbood, Zaiqiao Meng, and Nigel Collier. 2025. Time to Revisit Exact Match. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 11903–11926, Suzhou, China. Association for Computational Linguistics.