@inproceedings{huang-etal-2025-mre,
title = "{MRE}-{MI}: A Multi-image Dataset for Multimodal Relation Extraction in Social Media Posts",
author = "Huang, Shizhou and
Xu, Bo and
Li, Changqun and
Yu, Yang and
Lin, Xin Alex",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Findings of the Association for Computational Linguistics: NAACL 2025",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.351/",
pages = "6267--6277",
ISBN = "979-8-89176-195-7",
abstract = "Despite recent advances in Multimodal Relation Extraction (MRE), existing datasets and approaches primarily focus on single-image scenarios, overlooking the prevalent real-world cases where relationships are expressed through multiple images alongside text. To address this limitation, we present MRE-MI, a novel human-annotated dataset that includes both multi-image and single-image instances for relation extraction. Beyond dataset creation, we establish comprehensive baselines and propose a simple model named Global and Local Relevance-Modulated Attention Model (GLRA) to address the new challenges in multi-image scenarios. Our extensive experiments reveal that incorporating multiple images substantially improves relation extraction in multi-image scenarios. Furthermore, GLRA achieves state-of-the-art results on MRE-MI, demonstrating its effectiveness. The datasets and source code can be found at https://github.com/JinFish/MRE-MI."
}
Markdown (Informal)
[MRE-MI: A Multi-image Dataset for Multimodal Relation Extraction in Social Media Posts](https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.351/) (Huang et al., Findings 2025)
ACL