Abstract
This paper presents our approach for the 2024 ChemoTimelines shared task. Specifically, we explored using Large Language Models (LLMs) for temporal relation extraction. We evaluate multiple model variations based on how the training data is used. For instance, we transform the task into a question-answering problem and use QA pairs to extract chemo-related events and their temporal relations. Next, we add all the documents to each question-answer pair as examples in our training dataset. Finally, we explore adding unlabeled data for continued pretraining. Each addition is done iteratively. Our results show that adding the document helps, but unlabeled data does not yield performance improvements, possibly because we used only 1% of the available data. Moreover, we find that instruction-tuned models still substantially underperform more traditional systems (e.g., EntityBERT).- Anthology ID:
- 2024.clinicalnlp-1.58
- Volume:
- Proceedings of the 6th Clinical Natural Language Processing Workshop
- Month:
- June
- Year:
- 2024
- Address:
- Mexico City, Mexico
- Editors:
- Tristan Naumann, Asma Ben Abacha, Steven Bethard, Kirk Roberts, Danielle Bitterman
- Venues:
- ClinicalNLP | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 604–615
- Language:
- URL:
- https://aclanthology.org/2024.clinicalnlp-1.58
- DOI:
- 10.18653/v1/2024.clinicalnlp-1.58
- Cite (ACL):
- Xingmeng Zhao and Anthony Rios. 2024. UTSA-NLP at ChemoTimelines 2024: Evaluating Instruction-Tuned Language Models for Temporal Relation Extraction. In Proceedings of the 6th Clinical Natural Language Processing Workshop, pages 604–615, Mexico City, Mexico. Association for Computational Linguistics.
- Cite (Informal):
- UTSA-NLP at ChemoTimelines 2024: Evaluating Instruction-Tuned Language Models for Temporal Relation Extraction (Zhao & Rios, ClinicalNLP-WS 2024)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-4/2024.clinicalnlp-1.58.pdf