@inproceedings{kim-etal-2025-ko,
title = "{K}o-{L}ong{RAG}: A {K}orean Long-Context {RAG} Benchmark Built with a Retrieval-Free Approach",
author = "Kim, Yongil and
Yeen, Heuiyeen and
Yun, Hyeongu and
Lee, Jinsik",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.938/",
doi = "10.18653/v1/2025.findings-emnlp.938",
pages = "17317--17329",
ISBN = "979-8-89176-335-7",
abstract = "The rapid advancement of large language models (LLMs) significantly enhances long-context Retrieval-Augmented Generation (RAG), yet existing benchmarks focus primarily on English. This leaves low-resource languages without comprehensive evaluation frameworks, limiting their progress in retrieval-based tasks. To bridge this gap, we introduce Ko-LongRAG, the first Korean long-context RAG benchmark. Unlike conventional benchmarks that depend on external retrievers, Ko-LongRAG adopts a retrieval-free approach designed around Specialized Content Knowledge (SCK), enabling controlled and high-quality QA pair generation without the need for an extensive retrieval infrastructure. Our evaluation shows that o1 model achieves the highest performance among proprietary models, while EXAONE 3.5 leads among open-sourced models. Additionally, various findings confirm Ko-LongRAG as a reliable benchmark for assessing Korean long-context RAG capabilities and highlight its potential for advancing multilingual RAG research. The dataset and source code will be released publicly."
}Markdown (Informal)
[Ko-LongRAG: A Korean Long-Context RAG Benchmark Built with a Retrieval-Free Approach](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.938/) (Kim et al., Findings 2025)
ACL