@inproceedings{li-etal-2025-exploring-multilingual,
title = "Exploring Multilingual Probing in Large Language Models: A Cross-Language Analysis",
author = "Li, Daoyang and
Zhao, Haiyan and
Zeng, Qingcheng and
Du, Mengnan",
editor = "Fei, Hao and
Tu, Kewei and
Zhang, Yuhui and
Hu, Xiang and
Han, Wenjuan and
Jia, Zixia and
Zheng, Zilong and
Cao, Yixin and
Zhang, Meishan and
Lu, Wei and
Siddharth, N. and
{\O}vrelid, Lilja and
Xue, Nianwen and
Zhang, Yue",
booktitle = "Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025)",
month = aug,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.xllm-1.7/",
pages = "61--70",
ISBN = "979-8-89176-286-2",
abstract = "Probing techniques for large language models (LLMs) have primarily focused on English, overlooking the vast majority of other world{'}s languages. In this paper, we extend these probing methods to a multilingual context, investigating how LLMs encode linguistic structures across diverse languages. We conduct experiments on several open-source LLM models, analyzing probing accuracy, trends across layers, and similarities between probing vectors for multiple languages. Our key findings reveal: (1) a consistent performance gap between high-resource and low-resource languages, with high-resource languages achieving significantly higher probing accuracy; (2) divergent layer-wise accuracy trends, where high-resource languages show substantial improvement in deeper layers similar to English; and (3) higher representational similarities among high-resource languages, with low-resource languages demonstrating lower similarities both among themselves and with high-resource languages. These results provide insights into how linguistic structures are represented differently across languages in LLMs and emphasize the need for improved structure modeling for low-resource languages."
}
Markdown (Informal)
[Exploring Multilingual Probing in Large Language Models: A Cross-Language Analysis](https://preview.aclanthology.org/landing_page/2025.xllm-1.7/) (Li et al., XLLM 2025)
ACL