Ko-LongRAG: A Korean Long-Context RAG Benchmark Built with a Retrieval-Free Approach

Yongil Kim, Heuiyeen Yeen, Hyeongu Yun, Jinsik Lee


Abstract
The rapid advancement of large language models (LLMs) significantly enhances long-context Retrieval-Augmented Generation (RAG), yet existing benchmarks focus primarily on English. This leaves low-resource languages without comprehensive evaluation frameworks, limiting their progress in retrieval-based tasks. To bridge this gap, we introduce Ko-LongRAG, the first Korean long-context RAG benchmark. Unlike conventional benchmarks that depend on external retrievers, Ko-LongRAG adopts a retrieval-free approach designed around Specialized Content Knowledge (SCK), enabling controlled and high-quality QA pair generation without the need for an extensive retrieval infrastructure. Our evaluation shows that o1 model achieves the highest performance among proprietary models, while EXAONE 3.5 leads among open-sourced models. Additionally, various findings confirm Ko-LongRAG as a reliable benchmark for assessing Korean long-context RAG capabilities and highlight its potential for advancing multilingual RAG research. The dataset and source code will be released publicly.
Anthology ID:
2025.findings-emnlp.938
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
17317–17329
Language:
URL:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.938/
DOI:
10.18653/v1/2025.findings-emnlp.938
Bibkey:
Cite (ACL):
Yongil Kim, Heuiyeen Yeen, Hyeongu Yun, and Jinsik Lee. 2025. Ko-LongRAG: A Korean Long-Context RAG Benchmark Built with a Retrieval-Free Approach. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 17317–17329, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Ko-LongRAG: A Korean Long-Context RAG Benchmark Built with a Retrieval-Free Approach (Kim et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.938.pdf
Checklist:
 2025.findings-emnlp.938.checklist.pdf