@inproceedings{yu-etal-2024-lfed,
title = "{LFED}: A Literary Fiction Evaluation Dataset for Large Language Models",
author = "Yu, Linhao and
Liu, Qun and
Xiong, Deyi",
editor = "Calzolari, Nicoletta and
Kan, Min-Yen and
Hoste, Veronique and
Lenci, Alessandro and
Sakti, Sakriani and
Xue, Nianwen",
booktitle = "Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.lrec-main.915/",
pages = "10466--10475",
abstract = "The rapid evolution of large language models (LLMs) has ushered in the need for comprehensive assessments of their performance across various dimensions. In this paper, we propose LFED, a Literary Fiction Evaluation Dataset, which aims to evaluate the capability of LLMs on the long fiction comprehension and reasoning. We collect 95 literary fictions that are either originally written in Chinese or translated into Chinese, covering a wide range of topics across several centuries. We define a question taxonomy with 8 question categories to guide the creation of 1,304 questions. Additionally, we conduct an in-depth analysis to ascertain how specific attributes of literary fictions (e.g., novel types, character numbers, the year of publication) impact LLM performance in evaluations. Through a series of experiments involving various state-of-the-art LLMs, our findings reveal that these models face considerable challenges in effectively addressing questions related to literary fictions, with ChatGPT reaching only 57.08{\%} under the zero-shot setting. The dataset will be publicly available at https://github.com/tjunlp-lab/LFED.git."
}
Markdown (Informal)
[LFED: A Literary Fiction Evaluation Dataset for Large Language Models](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.lrec-main.915/) (Yu et al., LREC-COLING 2024)
ACL