@inproceedings{qiu-hu-2025-deep,
    title = "Deep Associations, High Creativity: A Simple yet Effective Metric for Evaluating Large Language Models",
    author = "Qiu, Ziliang  and
      Hu, Renfen",
    editor = "Christodoulopoulos, Christos  and
      Chakraborty, Tanmoy  and
      Rose, Carolyn  and
      Peng, Violet",
    booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.550/",
    pages = "10870--10883",
    ISBN = "979-8-89176-332-6",
    abstract = "The evaluation of LLMs' creativity represents a crucial research domain, though challenges such as data contamination and costly human assessments often impede progress. Drawing inspiration from human creativity assessment, we propose PACE, asking LLMs to generate Parallel Chains of Associations to Evaluate their creativity. PACE minimizes the risk of data contamination and offers a straightforward, highly efficient evaluation, as evidenced by its strong correlation with Arena Creative Writing (Spearman{'}s $\rho = 0.739$, $p < 0.001$) on various proprietary and open-source models. A comparative analysis of associative creativity between LLMs and humans reveals that while high-performing LLMs achieve scores comparable to average human performance, top-performing humans consistently outperform LLMs. Furthermore, linguistic analysis reveals that both humans and LLMs exhibit a trend of decreasing concreteness in their associations, and humans demonstrating a greater diversity of associative patterns."
}Markdown (Informal)
[Deep Associations, High Creativity: A Simple yet Effective Metric for Evaluating Large Language Models](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.550/) (Qiu & Hu, EMNLP 2025)
ACL