@inproceedings{ye-etal-2025-optimising,
title = "Optimising Factual Consistency in Summarisation via Preference Learning from Multiple Imperfect Metrics",
author = "Ye, Yuxuan and
Santos-Rodriguez, Raul and
Simpson, Edwin",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.940/",
doi = "10.18653/v1/2025.findings-emnlp.940",
pages = "17342--17355",
ISBN = "979-8-89176-335-7",
abstract = "Reinforcement learning with evaluation metrics as rewards is widely used to enhance specific capabilities of language models. However, for tasks such as factually consistent summarisation, existing metrics remain underdeveloped, limiting their effectiveness as signals for shaping model behaviour.While individual factuality metrics are unreliable, their combination can more effectively capture diverse factual errors. We leverage this insight to introduce an automated training pipeline that improves factual consistency in summaries by aggregating scores from different weak metrics. Our approach avoids the need for complex reward shaping by mapping scores to preferences and filtering out cases with high disagreement between metrics. For each source document, we generate lexically similar summary pairs by varying decoding strategies, enabling the model to learn from factual differences caused by subtle lexical differences. This approach constructs a high-quality preference dataset using only source documents.Experiments demonstrate consistent factuality gains across models, ranging from early encoder-decoder architectures to modern large language models, with smaller models reaching comparable factuality to larger ones."
}Markdown (Informal)
[Optimising Factual Consistency in Summarisation via Preference Learning from Multiple Imperfect Metrics](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.940/) (Ye et al., Findings 2025)
ACL