Chengzhi Liu
2025
MMRC: A Large-Scale Benchmark for Understanding Multimodal Large Language Model in Real-World Conversation
Haochen Xue
|
Feilong Tang
|
Ming Hu
|
Yexin Liu
|
Qidong Huang
|
Yulong Li
|
Chengzhi Liu
|
Zhongxing Xu
|
Chong Zhang
|
Chun-Mei Feng
|
Yutong Xie
|
Imran Razzak
|
Zongyuan Ge
|
Jionglong Su
|
Junjun He
|
Yu Qiao
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Recent multimodal large language models (MLLMs) have demonstrated significant potential in open-ended conversation, generating more accurate and personalized responses. However, their abilities to memorize, recall, and reason in sustained interactions within real-world scenarios remain underexplored. This paper introduces MMRC, a Multi-Modal Real-world Conversation benchmark for evaluating six core open-ended abilities of MLLMs: information extraction, multi-turn reasoning, information update, image management, memory recall, and answer refusal. With data collected from real-world scenarios, MMRC comprises 5,120 conversations and 28,720 corresponding manually labeled questions, posing a significant challenge to existing MLLMs. Evaluations on 20 MLLMs in MMRC indicate an accuracy drop during open-ended interactions. We identify four common failure patterns: long-term memory degradation, inadequacies in updating factual knowledge, accumulated assumption of error propagation, and reluctance to “say no.” To mitigate these issues, we propose a simple yet effective NOTE-TAKING strategy, which can record key information from the conversation and remind the model during its responses, enhancing conversational capabilities. Experiments across six MLLMs demonstrate significant performance improvements.
Search
Fix author
Co-authors
- Chun-Mei Feng 1
- Zongyuan Ge 1
- Junjun He 1
- Ming Hu 1
- Qidong Huang 1
- show all...
Venues
- acl1