@inproceedings{tran-kretchmar-2024-towards,
title = "Towards Robust Extractive Question Answering Models: Rethinking the Training Methodology",
author = "Tran, Son Quoc and
Kretchmar, Matt",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2024.findings-emnlp.121/",
doi = "10.18653/v1/2024.findings-emnlp.121",
pages = "2222--2236",
abstract = "This paper proposes a novel training method to improve the robustness of Extractive Question Answering (EQA) models. Previous research has shown that existing models, when trained on EQA datasets that include unanswerable questions, demonstrate a significant lack of robustness against distribution shifts and adversarial attacks. Despite this, the inclusion of unanswerable questions in EQA training datasets is essential for ensuring real-world reliability. Our proposed training method includes a novel loss function for the EQA problem and challenges an implicit assumption present in numerous EQA datasets. Models trained with our method maintain in-domain performance while achieving a notable improvement on out-of-domain datasets. This results in an overall F1 score improvement of 5.7 across all testing sets. Furthermore, our models exhibit significantly enhanced robustness against two types of adversarial attacks, with a performance decrease of only about one-third compared to the default models."
}
Markdown (Informal)
[Towards Robust Extractive Question Answering Models: Rethinking the Training Methodology](https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2024.findings-emnlp.121/) (Tran & Kretchmar, Findings 2024)
ACL