MRE-MI: A Multi-image Dataset for Multimodal Relation Extraction in Social Media Posts

Shizhou Huang, Bo Xu, Changqun Li, Yang Yu, Xin Alex Lin


Abstract
Despite recent advances in Multimodal Relation Extraction (MRE), existing datasets and approaches primarily focus on single-image scenarios, overlooking the prevalent real-world cases where relationships are expressed through multiple images alongside text. To address this limitation, we present MRE-MI, a novel human-annotated dataset that includes both multi-image and single-image instances for relation extraction. Beyond dataset creation, we establish comprehensive baselines and propose a simple model named Global and Local Relevance-Modulated Attention Model (GLRA) to address the new challenges in multi-image scenarios. Our extensive experiments reveal that incorporating multiple images substantially improves relation extraction in multi-image scenarios. Furthermore, GLRA achieves state-of-the-art results on MRE-MI, demonstrating its effectiveness. The datasets and source code can be found at https://github.com/JinFish/MRE-MI.
Anthology ID:
2025.findings-naacl.351
Volume:
Findings of the Association for Computational Linguistics: NAACL 2025
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6267–6277
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.351/
DOI:
Bibkey:
Cite (ACL):
Shizhou Huang, Bo Xu, Changqun Li, Yang Yu, and Xin Alex Lin. 2025. MRE-MI: A Multi-image Dataset for Multimodal Relation Extraction in Social Media Posts. In Findings of the Association for Computational Linguistics: NAACL 2025, pages 6267–6277, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
MRE-MI: A Multi-image Dataset for Multimodal Relation Extraction in Social Media Posts (Huang et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.351.pdf