Linking Knowledge to Care: Knowledge Graph-Augmented Medical Follow-Up Question Generation

Liwen Sun, Xiang Yu, Ming Tan, Zhuohao Chen, Anqi Cheng, Ashutosh Joshi, Chenyan Xiong


Abstract
Clinical diagnosis is time-consuming, requiring intensive interactions between patients and medical professionals. While large language models (LLMs) could ease the pre-diagnostic workload, their limited domain knowledge hinders effective medical question generation. We introduce a Knowledge Graph-augmented LLM with active in-context learning to generate relevant and important follow-up questions, KG-Followup, serving as a critical module for the pre-diagnostic assessment. The structured medical domain knowledge graph serves as a seamless patch-up to provide professional domain expertise upon which the LLM can reason. Experiments demonstrate that KG-Followup outperforms state-of-the-art methods by 5% - 8% on relevant benchmarks.
Anthology ID:
2026.findings-eacl.43
Volume:
Findings of the Association for Computational Linguistics: EACL 2026
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
846–853
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.43/
DOI:
Bibkey:
Cite (ACL):
Liwen Sun, Xiang Yu, Ming Tan, Zhuohao Chen, Anqi Cheng, Ashutosh Joshi, and Chenyan Xiong. 2026. Linking Knowledge to Care: Knowledge Graph-Augmented Medical Follow-Up Question Generation. In Findings of the Association for Computational Linguistics: EACL 2026, pages 846–853, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Linking Knowledge to Care: Knowledge Graph-Augmented Medical Follow-Up Question Generation (Sun et al., Findings 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.findings-eacl.43.pdf
Checklist:
 2026.findings-eacl.43.checklist.pdf