Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization
Kellen Tan Cheng, Ganesh Ramesh, Nafiul Rashid, Geoffrey Jay Tso, Jilong Kuang
Abstract
Health wellness agents typically rely on large language models (LLMs) for response generation, where contextual information from a user’s conversation history can be used for response grounding and personalization. High-quality conversation summaries are one such method which can reduce the number of input tokens during response generation, decreasing overhead and inference latency. However, directly purposing LLMs for this task is infeasible due to the scale of the task, the compute overhead, and health data compliance regulations. Furthermore, ground truth for real-world datasets is scarce due to privacy concerns and the high cost of health expert annotators. These factors necessitate the development of small, potentially on-device, language models capable of health dialogue summarization, particularly in the absence of ground truth labels. In this paper, we first present a comprehensive empirical study that benchmarks a variety of state-of-the-art smaller language models to better understand their baseline capabilities. Second, we present an unsupervised method that uses the summaries from multiple models, refined with iterative feedback, to generate high-quality summaries of health dialogues. Experiments illustrate that our method is able to outperform baseline on both open-source and proprietary benchmarks. Notably, our method can be run viably on local compute without a GPU, using just a single Macbook with 16 GB of memory.- Anthology ID:
- 2026.eacl-long.105
- Volume:
- Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
- Month:
- March
- Year:
- 2026
- Address:
- Rabat, Morocco
- Editors:
- Vera Demberg, Kentaro Inui, Lluís Marquez
- Venue:
- EACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 2340–2363
- Language:
- URL:
- https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.105/
- DOI:
- Cite (ACL):
- Kellen Tan Cheng, Ganesh Ramesh, Nafiul Rashid, Geoffrey Jay Tso, and Jilong Kuang. 2026. Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2340–2363, Rabat, Morocco. Association for Computational Linguistics.
- Cite (Informal):
- Compact Language Models with Iterative Text Refinement for Health Dialogue Summarization (Cheng et al., EACL 2026)
- PDF:
- https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.105.pdf