Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization

Kellen Tan Cheng, Ganesh Ramesh, Nafiul Rashid, Geoffrey Jay Tso, Jilong Kuang


Abstract
Health wellness agents typically rely on large language models (LLMs) for response generation, where contextual information from a user’s conversation history can be used for response grounding and personalization. High-quality conversation summaries are one such method which can reduce the number of input tokens during response generation, decreasing overhead and inference latency. However, directly purposing LLMs for this task is infeasible due to the scale of the task, the compute overhead, and health data compliance regulations. Furthermore, ground truth for real-world datasets is scarce due to privacy concerns and the high cost of health expert annotators. These factors necessitate the development of small, potentially on-device, language models capable of health dialogue summarization, particularly in the absence of ground truth labels. In this paper, we first present a comprehensive empirical study that benchmarks a variety of state-of-the-art smaller language models to better understand their baseline capabilities. Second, we present an unsupervised method that uses the summaries from multiple models, refined with iterative feedback, to generate high-quality summaries of health dialogues. Experiments illustrate that our method is able to outperform baseline on both open-source and proprietary benchmarks. Notably, our method can be run viably on local compute without a GPU, using just a single Macbook with 16 GB of memory.
Anthology ID:
2026.eacl-long.105
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2340–2363
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.105/
DOI:
Bibkey:
Cite (ACL):
Kellen Tan Cheng, Ganesh Ramesh, Nafiul Rashid, Geoffrey Jay Tso, and Jilong Kuang. 2026. Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2340–2363, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization (Cheng et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.105.pdf