Are Knowledge and Reference in Multilingual Language Models Cross-Lingually Consistent?

Xi Ai, Mahardika Krisna Ihsani, Min-Yen Kan


Abstract
Cross-lingual consistency should be considered to assess cross-lingual transferability, maintain the factuality of the model knowledge across languages, and preserve the parity of language model performance. We are thus interested in analyzing, evaluating, and interpreting cross-lingual consistency for factual knowledge.To facilitate our study, we examine multiple pretrained models and tuned models with code-mixed coreferential statements that convey identical knowledge across languages. Interpretability approaches are leveraged to analyze the behavior of a model in cross-lingual contexts, showing different levels of consistency in multilingual models, subject to language families, linguistic factors, scripts, and a bottleneck in cross-lingual consistency on a particular layer. Code-switching training and cross-lingual word alignment objectives show the most promising results, emphasizing the worthiness of cross-lingual alignment supervision and code-switching strategies for both multilingual performance and cross-lingual consistency enhancement. In addition, experimental results suggest promising result for calibrating consistency on test time via activation patching.
Anthology ID:
2025.findings-emnlp.267
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4975–5011
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.267/
DOI:
10.18653/v1/2025.findings-emnlp.267
Bibkey:
Cite (ACL):
Xi Ai, Mahardika Krisna Ihsani, and Min-Yen Kan. 2025. Are Knowledge and Reference in Multilingual Language Models Cross-Lingually Consistent?. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 4975–5011, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Are Knowledge and Reference in Multilingual Language Models Cross-Lingually Consistent? (Ai et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.267.pdf
Checklist:
 2025.findings-emnlp.267.checklist.pdf