PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality

Byeongho Yu, Changhun Lee, Jun-gyu Jin, Eunhyeok Park


Abstract
To mitigate the hallucination problem in large language models, DoLa exploits early exit logits from the same model as a contrastive prior. However, we found that these early exit logits tend to be flat, low in magnitude, and fail to reflect meaningful contrasts. To address this, we propose PruneCD, a novel contrastive decoding method that constructs the amateur model via layer pruning rather than early exit. This design leads to more informative and well-aligned logits, enabling more effective contrastive decoding. Through qualitative and quantitative analyses, we demonstrate that PruneCD consistently improves factuality with minimal inference overhead, offering a robust and practical approach to mitigating hallucinations in LLMs.
Anthology ID:
2025.emnlp-main.1651
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
32450–32461
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1651/
DOI:
Bibkey:
Cite (ACL):
Byeongho Yu, Changhun Lee, Jun-gyu Jin, and Eunhyeok Park. 2025. PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 32450–32461, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality (Yu et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1651.pdf
Checklist:
 2025.emnlp-main.1651.checklist.pdf