Improving Pre-trained Language Models with Knowledge Enhancement and Filtering Framework
Qi Zhao, Qi Song, Tian Xie, Haiyue Zhang, Hongyu Yang, Xiangyang Li
Abstract
Pre-trained language models (PLMs) are widely used in NLP but struggle with capturing entity knowledge. To address this, knowledge enhancement techniques have been proposed. However, existing methods rely heavily on external knowledge bases embedding and often introduce noisy entity representations. In this work, we propose a novel **K**nowledge **E**nhancement **F**iltering **F**ramework named KEFF, which contains both knowledge enhancement and knowledge enhancement filtering modules for PLM. We find that there are certain redundant bits in the embedding space of PLMs. Building on this insight, we implement knowledge-enhanced mapping of redundant bit values in entity span tokens. In order to solve the knowledge enhancement problem of existing methods that introduce noisy entity representation knowledge, we further propose a novel knowledge enhancement filter based on our knowledge enhancement method. Finally, experiments on four knowledge-driven NLP tasks show that our method effectively improves the ability of PLMs on downstream tasks. Compared to state-of-the-art approachs, our method achieves the highest F1-score and accuracy, while reducing the computational cost by 1.7-2.5x.- Anthology ID:
- 2025.findings-naacl.213
- Volume:
- Findings of the Association for Computational Linguistics: NAACL 2025
- Month:
- April
- Year:
- 2025
- Address:
- Albuquerque, New Mexico
- Editors:
- Luis Chiruzzo, Alan Ritter, Lu Wang
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 3860–3871
- Language:
- URL:
- https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.213/
- DOI:
- Cite (ACL):
- Qi Zhao, Qi Song, Tian Xie, Haiyue Zhang, Hongyu Yang, and Xiangyang Li. 2025. Improving Pre-trained Language Models with Knowledge Enhancement and Filtering Framework. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 3860–3871, Albuquerque, New Mexico. Association for Computational Linguistics.
- Cite (Informal):
- Improving Pre-trained Language Models with Knowledge Enhancement and Filtering Framework (Zhao et al., Findings 2025)
- PDF:
- https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.213.pdf