Efficient Medical Question Answering with Knowledge-Augmented Question Generation

Julien Khlaut, Corentin Dancette, Elodie Ferreres, Benani Alaedine, Herent Herent, Pierre Manceron


Abstract
In the expanding field of language model applications, medical knowledge representation remains a significant challenge due to the specialized nature of the domain. Large language models, such as GPT-4, obtain reasonable scores on medical question-answering tasks, but smaller models are far behind.In this work, we introduce a method to improve the proficiency of a small language model in the medical domain by employing a two-fold approach. We first fine-tune the model on a corpus of medical textbooks. Then, we use GPT-4 to generate questions similar to the downstream task, prompted with textbook knowledge, and use them to fine-tune the model. Additionally, we introduce ECN-QA, a novel Medical QA dataset containing “progressive questions” composed of related sequential questions. We show the benefits of our training strategy on this dataset.The study’s findings highlight the potential of small language models in the medical domain when appropriately fine-tuned.
Anthology ID:
2024.clinicalnlp-1.2
Volume:
Proceedings of the 6th Clinical Natural Language Processing Workshop
Month:
June
Year:
2024
Address:
Mexico City, Mexico
Editors:
Tristan Naumann, Asma Ben Abacha, Steven Bethard, Kirk Roberts, Danielle Bitterman
Venues:
ClinicalNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
10–20
Language:
URL:
https://aclanthology.org/2024.clinicalnlp-1.2
DOI:
Bibkey:
Cite (ACL):
Julien Khlaut, Corentin Dancette, Elodie Ferreres, Benani Alaedine, Herent Herent, and Pierre Manceron. 2024. Efficient Medical Question Answering with Knowledge-Augmented Question Generation. In Proceedings of the 6th Clinical Natural Language Processing Workshop, pages 10–20, Mexico City, Mexico. Association for Computational Linguistics.
Cite (Informal):
Efficient Medical Question Answering with Knowledge-Augmented Question Generation (Khlaut et al., ClinicalNLP-WS 2024)
Copy Citation:
PDF:
https://preview.aclanthology.org/jeptaln-2024-ingestion/2024.clinicalnlp-1.2.pdf