@inproceedings{you-etal-2025-pali,
title = "{PALI}-{NLP} at {S}em{E}val 2025 Task 1: Multimodal Idiom Recognition and Alignment",
author = "You, Runyang and
Mei, Xinyue and
Zhou, Mengyuan",
editor = "Rosenthal, Sara and
Ros{\'a}, Aiala and
Ghosh, Debanjan and
Zampieri, Marcos",
booktitle = "Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.161/",
pages = "1211--1216",
ISBN = "979-8-89176-273-2",
abstract = "Understanding idioms in multimodal contexts poses significant challenges due to data scarcity, idiomatic ambiguity, and the need for effective alignment of visual and textual inputs. In this work, we introduce MIRA (Multimodal Idiom Recognition and Alignment), a training-free framework designed to address these challenges on the SemEval-2025 Task 1 (AdMIRe) benchmark. MIRA leverages powerful closed-source large language models (LLMs) and integrates three key innovations: bias correction via in-context learning, multi-step semantic-visual fusion, and a self-revision mechanism that iteratively refines its outputs through backward verification. By systematically processing and fusing multimodal inputs, MIRA generates high-quality, fine-grained image-text representations that enhance idiom comprehension across different languages and cultural contexts. Experimental evaluations in both English and Portuguese demonstrate that our approach achieves robust performance without the need for additional training, setting a new standard for multimodal idiom recognition."
}
Markdown (Informal)
[PALI-NLP at SemEval 2025 Task 1: Multimodal Idiom Recognition and Alignment](https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.161/) (You et al., SemEval 2025)
ACL