Document-Level Event-Argument Data Augmentation for Challenging Role Types
Joseph Gatto, Omar Sharif, Parker Seegmiller, Sarah Masud Preum
Abstract
Event Argument Extraction (EAE) is a daunting information extraction problem — with significant limitations in few-shot cross-domain (FSCD) settings. A common solution to FSCD modeling is data augmentation. Unfortunately, existing augmentation methods are not well-suited to a variety of real-world EAE contexts, including (i) modeling long documents (documents with over 10 sentences), and (ii) modeling challenging role types (i.e., event roles with little to no training data and semantically outlying roles). We introduce two novel LLM-powered data augmentation methods for generating extractive document-level EAE samples using zero in-domain training data. We validate the generalizability of our approach on four datasets — showing significant performance increases in low-resource settings. Our highest performing models provide a 13-pt increase in F1 score on zero-shot role extraction in FSCD evaluation.- Anthology ID:
- 2025.acl-long.1221
- Volume:
- Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
- Month:
- July
- Year:
- 2025
- Address:
- Vienna, Austria
- Editors:
- Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
- Venue:
- ACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 25109–25131
- Language:
- URL:
- https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1221/
- DOI:
- Cite (ACL):
- Joseph Gatto, Omar Sharif, Parker Seegmiller, and Sarah Masud Preum. 2025. Document-Level Event-Argument Data Augmentation for Challenging Role Types. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 25109–25131, Vienna, Austria. Association for Computational Linguistics.
- Cite (Informal):
- Document-Level Event-Argument Data Augmentation for Challenging Role Types (Gatto et al., ACL 2025)
- PDF:
- https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1221.pdf