bnContextQA: Benchmarking Long-Context Question Answering and Challenges in Bangla
Adnan Ahmad, Labiba Adiba, Namirah Rasul, Md Tahmid Rahman Laskar, Sabbir Ahmed
Abstract
Large models have advanced in processing long input sequences, but their ability to consistently use information across extended contexts remains a challenge. Recent studies highlight a positional bias where models prioritize information at the beginning or end of the input while neglecting the middle, resulting in a U-shaped performance curve but this was limited to English. Whether this bias is universal or shaped by language-specific factors remains unclear. In this work, we investigate positional bias in Bangla, a widely spoken but computationally underrepresented language. To support this, we introduce a novel Bangla benchmark dataset, bnContextQA, specifically designed for long-context comprehension. The dataset comprises of 350 long-context QA instances, each paired with 30 context paragraphs, allowing controlled evaluation of information retrieval at different positions. Using this dataset, we assess the performance of LLMs on Bangla across varying passage positions, providing insights into cross-linguistic positional effects. The bnContextQA dataset is publicly available at https://github.com/labiba02/bnContextQA.git to support future research on long-context understanding in Bangla and multilingual LLMs.- Anthology ID:
- 2025.banglalp-1.29
- Volume:
- Proceedings of the Second Workshop on Bangla Language Processing (BLP-2025)
- Month:
- December
- Year:
- 2025
- Address:
- Mumbai, India
- Editors:
- Firoj Alam, Sudipta Kar, Shammur Absar Chowdhury, Naeemul Hassan, Enamul Hoque Prince, Mohiuddin Tasnim, Md Rashad Al Hasan Rony, Md Tahmid Rahman Rahman
- Venues:
- BanglaLP | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 357–365
- Language:
- URL:
- https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.banglalp-1.29/
- DOI:
- Cite (ACL):
- Adnan Ahmad, Labiba Adiba, Namirah Rasul, Md Tahmid Rahman Laskar, and Sabbir Ahmed. 2025. bnContextQA: Benchmarking Long-Context Question Answering and Challenges in Bangla. In Proceedings of the Second Workshop on Bangla Language Processing (BLP-2025), pages 357–365, Mumbai, India. Association for Computational Linguistics.
- Cite (Informal):
- bnContextQA: Benchmarking Long-Context Question Answering and Challenges in Bangla (Ahmad et al., BanglaLP 2025)
- PDF:
- https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.banglalp-1.29.pdf