Do Multilingual Transformers Encode Paninian Grammatical Relations? A Layer-wise Probing Study

Akshit Kumar, Dipti Sharma, Parameswari Krishnamurthy


Abstract
Large multilingual transformers such as XLM-RoBERTa achieve impressive performance on diverse NLP benchmarks, but understanding how they internally encode grammatical information remains challenging. This study investigates the encoding of syntactic and morphological information derived from the Paninian grammatical framework—specifically designed for morphologically rich Indian languages—across model layers. Using diagnostic probing, we analyze the hidden representations of frozen XLM-RoBERTa-base, mBERT, and IndicBERT models across seven Indian languages (Hindi, Kannada, Malayalam, Marathi, Telugu, Urdu, Bengali). Probes are trained to predict Paninian dependency relations (by edge probing) and essential morphosyntactic features (UPOS tags, Vibhakti markers). We find that syntactic structure (dependencies) is primarily encoded in the middle-to-upper-middle layers (layers 6–9), while lexical features peak slightly earlier. Although the general layer-wise trends are shared across models, significant variations in absolute probing performance reflect differences in model capacity, pre-training data, and language-specific characteristics. These findings shed light on how theory-specific grammatical information emerges implicitly within multilingual transformer representations trained largely on unstructured raw text.
Anthology ID:
2025.quasy-1.15
Volume:
Proceedings of the Third Workshop on Quantitative Syntax (QUASY, SyntaxFest 2025)
Month:
August
Year:
2025
Address:
Ljubljana, Slovenia
Editors:
Xinying Chen, Yaqin Wang
Venues:
Quasy | WS | SyntaxFest
SIG:
SIGPARSE
Publisher:
Association for Computational Linguistics
Note:
Pages:
124–130
Language:
URL:
https://preview.aclanthology.org/transition-to-people-yaml/2025.quasy-1.15/
DOI:
Bibkey:
Cite (ACL):
Akshit Kumar, Dipti Sharma, and Parameswari Krishnamurthy. 2025. Do Multilingual Transformers Encode Paninian Grammatical Relations? A Layer-wise Probing Study. In Proceedings of the Third Workshop on Quantitative Syntax (QUASY, SyntaxFest 2025), pages 124–130, Ljubljana, Slovenia. Association for Computational Linguistics.
Cite (Informal):
Do Multilingual Transformers Encode Paninian Grammatical Relations? A Layer-wise Probing Study (Kumar et al., Quasy-SyntaxFest 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/transition-to-people-yaml/2025.quasy-1.15.pdf