Beyond Literal Meaning: How LLMs Interpret Yemeni Proverbs

Nasser Thmer, Ali Al-Laith, Muhammad Shoaib


Abstract
We present a benchmark Yemeni proverbs dataset paired with expert-annotated explanations, designed to evaluate the cultural reasoning abilities of large language models (LLMs). Using zero-shot and few-shot prompting, we assess seven LLMs through both automatic and human evaluation. Results show that instruction-tuned models like GPT-4o and Gemini 1.5 Pro outperform smaller models in both automatic and human evaluations. Few-shot prompting significantly improves performance across all models, underscoring its value for figurative and culturally grounded language tasks. Notably, ALLaM, a bilingual model trained on Arabic and English, achieves competitive results, demonstrating the potential of regionally adapted models for low-resource cultural tasks. LLM-as-a-Judge evaluation correlates strongly with human assessment (Kendall’s τ up to 0.98). Error analysis identifies recurring literal interpretation and cultural misalignment as key failure modes.
Anthology ID:
2026.lrec-main.83
Volume:
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Month:
May
Year:
2026
Address:
Palma de Mallorca, Spain
Editors:
Stelios Piperidis, Núria Bel, Henk van den Heuvel, Nancy Ide, Simon Krek, Antonio Toral
Venue:
LREC
SIG:
Publisher:
ELRA Language Resource Association
Note:
Pages:
1071–1080
Language:
URL:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.83/
DOI:
Bibkey:
Cite (ACL):
Nasser Thmer, Ali Al-Laith, and Muhammad Shoaib. 2026. Beyond Literal Meaning: How LLMs Interpret Yemeni Proverbs. International Conference on Language Resources and Evaluation, main:1071–1080.
Cite (Informal):
Beyond Literal Meaning: How LLMs Interpret Yemeni Proverbs (Thmer et al., LREC 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.83.pdf