CoCo-CoLa: Evaluating and Improving Language Adherence in Multilingual LLMs

Elnaz Rahmati, Alireza Salkhordeh Ziabari, Morteza Dehghani


Abstract
Multilingual Large Language Models (LLMs) develop cross-lingual abilities despite being trained on limited parallel data. However, they often struggle to generate responses in the intended language, favoring high-resource languages such as English. In this work, we introduce CoCo-CoLa (Correct Concept - Correct Language), a novel metric to evaluate language adherence in multilingual LLMs. Using fine-tuning experiments on a closed-book QA task across seven languages, we analyze how training in one language affects others’ performance. Our findings reveal that multilingual models share task knowledge across languages but exhibit biases in the selection of output language. We identify language-specific layers, showing that final layers play a crucial role in determining output language. Accordingly, we propose a partial training strategy that selectively fine-tunes key layers, improving language adherence while reducing computational cost. Our method achieves comparable or superior performance to full fine-tuning, particularly for low-resource languages, offering a more efficient multilingual adaptation.
Anthology ID:
2025.mrl-main.5
Volume:
Proceedings of the 5th Workshop on Multilingual Representation Learning (MRL 2025)
Month:
November
Year:
2025
Address:
Suzhuo, China
Editors:
David Ifeoluwa Adelani, Catherine Arnett, Duygu Ataman, Tyler A. Chang, Hila Gonen, Rahul Raja, Fabian Schmidt, David Stap, Jiayi Wang
Venues:
MRL | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
62–77
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.mrl-main.5/
DOI:
Bibkey:
Cite (ACL):
Elnaz Rahmati, Alireza Salkhordeh Ziabari, and Morteza Dehghani. 2025. CoCo-CoLa: Evaluating and Improving Language Adherence in Multilingual LLMs. In Proceedings of the 5th Workshop on Multilingual Representation Learning (MRL 2025), pages 62–77, Suzhuo, China. Association for Computational Linguistics.
Cite (Informal):
CoCo-CoLa: Evaluating and Improving Language Adherence in Multilingual LLMs (Rahmati et al., MRL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.mrl-main.5.pdf