Exploring the Generalizability of Factual Hallucination Mitigation via Enhancing Precise Knowledge Utilization

Siyuan Zhang, Yichi Zhang, Yinpeng Dong, Hang Su


Abstract
Large Language Models (LLMs) often struggle to align their responses with objective facts, resulting in the issue of factual hallucinations, which can be difficult to detect and mislead users without relevant knowledge. Although post-training techniques have been employed to mitigate the issue, existing methods usually suffer from poor generalization and trade-offs in other different capabilities. In this paper, we propose to address these by directly augmenting LLM’s fundamental ability to precisely leverage its knowledge and introduce PKUE (Precise Knowledge Utilization Enhancement), which fine-tunes the model on self-generated responses to precise and simple factual questions through preference optimization. Furthermore, we construct FactualBench, a comprehensive and precise factual QA dataset containing 181k Chinese data spanning 21 domains, to facilitate both evaluation and training. Extensive experiments demonstrate that PKUE significantly improves LLM overall performance, with consistent enhancement across factual tasks of various forms, general tasks beyond factuality, and tasks in different language.
Anthology ID:
2025.findings-emnlp.211
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3936–3968
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.211/
DOI:
10.18653/v1/2025.findings-emnlp.211
Bibkey:
Cite (ACL):
Siyuan Zhang, Yichi Zhang, Yinpeng Dong, and Hang Su. 2025. Exploring the Generalizability of Factual Hallucination Mitigation via Enhancing Precise Knowledge Utilization. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 3936–3968, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Exploring the Generalizability of Factual Hallucination Mitigation via Enhancing Precise Knowledge Utilization (Zhang et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.211.pdf
Checklist:
 2025.findings-emnlp.211.checklist.pdf