MATA: Mindful Assessment of the Telugu Abilities of Large Language Models

Chalamalasetti Kranti, Sowmya Vajjala


Abstract
In this paper, we introduce MATA, a novel evaluation dataset to assess the ability of Large Language Models (LLMs) in Telugu language, comprising 729 carefully curated multiple-choice and open-ended questions that span diverse linguistic dimensions. We evaluate 11 open-weight and closed-source LLMs on our dataset and present a fine-grained analysis of their performance. Further, we empirically show how LLMs rely on superficial heuristics such as answer position and distractor patterns for multiple-choice questions. Finally, we also compare LLM-as-a-judge evaluation with human evaluation for open-ended questions assess its reliability in a low-resource language. We argue that such fine-grained evaluation is essential for understanding model limitations and can inform the development of more linguistically capable LLMs, while also serving as a foundation for future research in Telugu NLP. Our dataset is available at:https://huggingface.co/datasets/TeluguLLMResearch/MATA
Anthology ID:
2026.lrec-main.334
Volume:
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Month:
May
Year:
2026
Address:
Palma de Mallorca, Spain
Editors:
Stelios Piperidis, Núria Bel, Henk van den Heuvel, Nancy Ide, Simon Krek, Antonio Toral
Venue:
LREC
SIG:
Publisher:
ELRA Language Resource Association
Note:
Pages:
4239–4256
Language:
URL:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.334/
DOI:
Bibkey:
Cite (ACL):
Chalamalasetti Kranti and Sowmya Vajjala. 2026. MATA: Mindful Assessment of the Telugu Abilities of Large Language Models. International Conference on Language Resources and Evaluation, main:4239–4256.
Cite (Informal):
MATA: Mindful Assessment of the Telugu Abilities of Large Language Models (Kranti & Vajjala, LREC 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.334.pdf