UTSA-NLP at ArchEHR-QA 2025: Improving EHR Question Answering via Self-Consistency Prompting

Sara Shields-Menard, Zach Reimers, Joshua Gardner, David Perry, Anthony Rios


Abstract
We describe our system for the ArchEHR-QA Shared Task on answering clinical questions using electronic health records (EHRs). Our approach uses large language models in two steps: first, to find sentences in the EHR relevant to a clinician’s question, and second, to generate a short, citation-supported response based on those sentences. We use few-shot prompting, self-consistency, and thresholding to improve the sentence classification step to decide which sentences are essential. We compare several models and find that a smaller 8B model performs better than a larger 70B model for identifying relevant information. Our results show that accurate sentence selection is critical for generating high-quality responses and that self-consistency with thresholding helps make these decisions more reliable.
Anthology ID:
2025.bionlp-share.10
Volume:
Proceedings of the 24th Workshop on Biomedical Language Processing (Shared Tasks)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Sarvesh Soni, Dina Demner-Fushman
Venues:
BioNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
81–90
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.bionlp-share.10/
DOI:
10.18653/v1/2025.bionlp-share.10
Bibkey:
Cite (ACL):
Sara Shields-Menard, Zach Reimers, Joshua Gardner, David Perry, and Anthony Rios. 2025. UTSA-NLP at ArchEHR-QA 2025: Improving EHR Question Answering via Self-Consistency Prompting. In Proceedings of the 24th Workshop on Biomedical Language Processing (Shared Tasks), pages 81–90, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
UTSA-NLP at ArchEHR-QA 2025: Improving EHR Question Answering via Self-Consistency Prompting (Shields-Menard et al., BioNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.bionlp-share.10.pdf