@inproceedings{kumar-etal-2025-multilingual,
title = "Do Multilingual Transformers Encode {P}aninian Grammatical Relations? A Layer-wise Probing Study",
author = "Kumar, Akshit and
Sharma, Dipti and
Krishnamurthy, Parameswari",
editor = "Chen, Xinying and
Wang, Yaqin",
booktitle = "Proceedings of the Third Workshop on Quantitative Syntax (QUASY, SyntaxFest 2025)",
month = aug,
year = "2025",
address = "Ljubljana, Slovenia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/transition-to-people-yaml/2025.quasy-1.15/",
pages = "124--130",
ISBN = "979-8-89176-293-0",
abstract = "Large multilingual transformers such as XLM-RoBERTa achieve impressive performance on diverse NLP benchmarks, but understanding how they internally encode grammatical information remains challenging. This study investigates the encoding of syntactic and morphological information derived from the Paninian grammatical framework{---}specifically designed for morphologically rich Indian languages{---}across model layers. Using diagnostic probing, we analyze the hidden representations of frozen XLM-RoBERTa-base, mBERT, and IndicBERT models across seven Indian languages (Hindi, Kannada, Malayalam, Marathi, Telugu, Urdu, Bengali). Probes are trained to predict Paninian dependency relations (by edge probing) and essential morphosyntactic features (UPOS tags, Vibhakti markers). We find that syntactic structure (dependencies) is primarily encoded in the middle-to-upper-middle layers (layers 6{--}9), while lexical features peak slightly earlier. Although the general layer-wise trends are shared across models, significant variations in absolute probing performance reflect differences in model capacity, pre-training data, and language-specific characteristics. These findings shed light on how theory-specific grammatical information emerges implicitly within multilingual transformer representations trained largely on unstructured raw text."
}
Markdown (Informal)
[Do Multilingual Transformers Encode Paninian Grammatical Relations? A Layer-wise Probing Study](https://preview.aclanthology.org/transition-to-people-yaml/2025.quasy-1.15/) (Kumar et al., Quasy-SyntaxFest 2025)
ACL