@inproceedings{hu-etal-2025-mcitebench,
title = "{MC}ite{B}ench: A Multimodal Benchmark for Generating Text with Citations",
author = "Hu, Caiyu and
Zhang, Yikai and
Zhu, Tinghui and
Ye, Yiwei and
Xiao, Yanghua",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.318/",
doi = "10.18653/v1/2025.findings-emnlp.318",
pages = "5949--5966",
ISBN = "979-8-89176-335-7",
abstract = "Multimodal Large Language Models (MLLMs) have advanced in integrating diverse modalities but frequently suffer from hallucination. A promising solution to mitigate this issue is to generate text with citations, providing a transparent chain for verification. However, existing work primarily focuses on generating citations for text-only content, leaving the challenges of multimodal scenarios largely unexplored. In this paper, we introduce MCiteBench, the first benchmark designed to assess the ability of MLLMs to generate text with citations in multimodal contexts. Our benchmark comprises data derived from academic papers and review-rebuttal interactions, featuring diverse information sources and multimodal content. Experimental results reveal that MLLMs struggle to ground their outputs reliably when handling multimodal input. Further analysis uncovers a systematic modality bias and reveals how models internally rely on different sources when generating citations, offering insights into model behavior and guiding future directions for multimodal citation tasks."
}Markdown (Informal)
[MCiteBench: A Multimodal Benchmark for Generating Text with Citations](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.318/) (Hu et al., Findings 2025)
ACL