Regularized Contrastive Decoding with Hard Negative Samples for LLM Hallucination Mitigation

Haonan Sheng, Dou Hu, Lingwei Wei, Wei Zhou, Songlin Hu


Abstract
Large language models are prone to generate hallucinations, which can undermine their reliability in high-stakes applications. Some works on LLM hallucination mitigation use the model’s internal signals to contrast different output during inference stage. However, these works often focus on simple forms of hallucinations, and struggle to effectively mitigate hallucinations. To address the issue, this paper exploits hard negative samples to construct a factually weaker model for improving contrastive decoding. We propose a new inference-time method, Regularized Contrastive Decoding (RCD), to capture correct hallucination signals for mitigating hallucinations in LLMs. RCD learns more diverse hallucination patterns via adversarial-aware fine-tuning and mitigates hallucinations via contrastive decoding. Experiments on four hallucination benchmarks demonstrate that our method achieves better LLM hallucination mitigation performance. Further analysis shows RCD generalizes well across different model sizes, task formats, perturbation methods and training data sizes.
Anthology ID:
2025.findings-emnlp.322
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6061–6073
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.322/
DOI:
10.18653/v1/2025.findings-emnlp.322
Bibkey:
Cite (ACL):
Haonan Sheng, Dou Hu, Lingwei Wei, Wei Zhou, and Songlin Hu. 2025. Regularized Contrastive Decoding with Hard Negative Samples for LLM Hallucination Mitigation. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 6061–6073, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Regularized Contrastive Decoding with Hard Negative Samples for LLM Hallucination Mitigation (Sheng et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.322.pdf
Checklist:
 2025.findings-emnlp.322.checklist.pdf