@inproceedings{zhang-etal-2024-need,
title = "Do We Need Language-Specific Fact-Checking Models? The Case of {C}hinese",
author = "Zhang, Caiqi and
Guo, Zhijiang and
Vlachos, Andreas",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.113/",
doi = "10.18653/v1/2024.emnlp-main.113",
pages = "1899--1914",
abstract = "This paper investigates the potential benefits of language-specific fact-checking models, focusing on the case of Chinese using CHEF dataset. To better reflect real-world fact-checking, we first develop a novel Chinese document-level evidence retriever, achieving state-of-the-art performance. We then demonstrate the limitations of translation-based methods and multilingual language models, highlighting the need for language-specific systems. To better analyze token-level biases in different systems, we construct an adversarial dataset based on the CHEF dataset, where each instance has a large word overlap with the original one but holds the opposite veracity label. Experimental results on the CHEF dataset and our adversarial dataset show that our proposed method outperforms translation-based methods and multilingual language models and is more robust toward biases, emphasizing the importance of language-specific fact-checking systems."
}
Markdown (Informal)
[Do We Need Language-Specific Fact-Checking Models? The Case of Chinese](https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.emnlp-main.113/) (Zhang et al., EMNLP 2024)
ACL