HaoYuan Ma


2025

pdf bib
Exploring Knowledge Filtering for Retrieval-Augmented Discriminative Tasks
Minjie Qiang | Zhongqing Wang | Xiaoyi Bao | HaoYuan Ma | Shoushan Li | Guodong Zhou
Findings of the Association for Computational Linguistics: ACL 2025

Retrieval-augmented methods have achieved remarkable advancements in alleviating the hallucination of large language models.Nevertheless, the introduction of external knowledge does not always lead to the expected improvement in model performance, as irrelevant or harmful information present in the retrieved knowledge can compromise the prediction process.To address these challenges, we propose a novel framework aimed at improving model performance by incorporating knowledge filtering and prediction fusion mechanisms.In particular, our approach first employs a perplexity-based annotation method to collect training data.Then, we design four distinct strategies to filter out harmful retrieved knowledge.Finally, we integrate the filtered knowledge to generate the final result via batch-wise predictions.We conduct extensive experiments across multiple discriminative task datasets to evaluate the proposed framework.The results demonstrate that our framework can significantly enhance the performance of models on discriminative tasks.