@inproceedings{lin-etal-2025-dynaquest,
title = "{D}yna{Q}uest: A Dynamic Question Answering Dataset Reflecting Real-World Knowledge Updates",
author = "Lin, Qian and
Li, Junyi and
Ng, Hwee Tou",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.1380/",
pages = "26918--26936",
ISBN = "979-8-89176-256-5",
abstract = "The rapidly changing nature of real-world information presents challenges for large language models (LLMs), which are typically trained on static datasets. This limitation makes it difficult for LLMs to accurately perform tasks that require up-to-date knowledge, such as time-sensitive question answering (QA). In this paper, we introduce **DynaQuest**, a **Dyna**mic **Quest**ion answering dataset reflecting knowledge updates in the real world. DynaQuest is based on Wikipedia Infoboxes, which are frequently updated to reflect real-world changes. Our dataset is created by automatically identifying and comparing changes between different versions of Wikipedia pages and generating question-answer pairs based on these updates. To address the challenges posed by our dynamic dataset, we propose **CARL**, a **C**ontext-**A**ware **R**einforcement **L**earning framework to improve the performance of LLMs on time-sensitive question answering. We conduct experiments on our collected dataset across recent time periods and demonstrate the effectiveness of our approach. Furthermore, we maintain a dynamic knowledge updating process, providing a periodically evolving benchmark to continually evaluate LLMs' ability to answer time-sensitive questions."
}
Markdown (Informal)
[DynaQuest: A Dynamic Question Answering Dataset Reflecting Real-World Knowledge Updates](https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.1380/) (Lin et al., Findings 2025)
ACL