RusConText Benchmark: A Russian Language Evaluation Benchmark for Understanding Context
Andrey Chirkin, Svetlana Kuznetsova, Maria Volina, Anna Dengina
Abstract
This paper represents an implementation of an approach rather similar to that of (Zhu et al., 2024), adapted for the Russian-language data. We introduce the RusConText Benchmark for evaluating short-context understanding in Russian, comprising four distinct yet interrelated tasks: coreference resolution, discourse understanding, idiom interpretation and ellipsis resolution. Each task targets a specific aspect of linguistic processing, challenging a large language model to recover omitted information, resolve referential dependencies, interpret idioms and discourse. The RusConText Benchmark is an additional resource beyond standard benchmarks, designed to assess model performance from a specific perspective. In addition, we present the results of scoring 4 models on our benchmark.- Anthology ID:
- 2025.acl-srw.91
- Volume:
- Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)
- Month:
- July
- Year:
- 2025
- Address:
- Vienna, Austria
- Editors:
- Jin Zhao, Mingyang Wang, Zhu Liu
- Venues:
- ACL | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 1158–1170
- Language:
- URL:
- https://preview.aclanthology.org/landing_page/2025.acl-srw.91/
- DOI:
- Cite (ACL):
- Andrey Chirkin, Svetlana Kuznetsova, Maria Volina, and Anna Dengina. 2025. RusConText Benchmark: A Russian Language Evaluation Benchmark for Understanding Context. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop), pages 1158–1170, Vienna, Austria. Association for Computational Linguistics.
- Cite (Informal):
- RusConText Benchmark: A Russian Language Evaluation Benchmark for Understanding Context (Chirkin et al., ACL 2025)
- PDF:
- https://preview.aclanthology.org/landing_page/2025.acl-srw.91.pdf