LIMICS at ArchEHR-QA 2025: Prompting LLMs Beats Fine-Tuned Embeddings

Adam Remaki, Armand Violle, Vikram Natraj, Étienne Guével, Akram Redjdal


Abstract
In this paper, we investigated two approaches to clinical question-answering based on patient-formulated questions, supported by their narratives and brief medical records. The first approach leverages zero- and few-shot prompt engineering techniques with GPT-based Large Language Models (LLMs), incorporating strategies such as prompt chaining and chain-of-thought reasoning to guide the models in generating answers. The second approach adopts a two-steps structure: first, a text-classification stage uses embedding-based models (e.g., BERT variants) to identify sentences within the medical record that are most relevant to the given question; then, we prompt an LLM to paraphrase them into an answer so that it is generated exclusively from these selected sentences. Our empirical results demonstrate that the first approach outperforms the classification-guided pipeline, achieving the highest score on the development set and the test set using prompt chaining. Code: github.com/armandviolle/BioNLP-2025
Anthology ID:
2025.bionlp-share.18
Volume:
BioNLP 2025 Shared Tasks
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Sarvesh Soni, Dina Demner-Fushman
Venues:
BioNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
150–159
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bionlp-share.18/
DOI:
Bibkey:
Cite (ACL):
Adam Remaki, Armand Violle, Vikram Natraj, Étienne Guével, and Akram Redjdal. 2025. LIMICS at ArchEHR-QA 2025: Prompting LLMs Beats Fine-Tuned Embeddings. In BioNLP 2025 Shared Tasks, pages 150–159, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
LIMICS at ArchEHR-QA 2025: Prompting LLMs Beats Fine-Tuned Embeddings (Remaki et al., BioNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.bionlp-share.18.pdf