LST at MWE-2026 AdMIRe 2: Advancing Multimodal Idiomaticity Representation

Le Qiu, Yu-Yin Hsu, Emmanuele Chersoni


Abstract
This paper presents our methods for the AdMIRe 2.0 shared task, which addresses multilingual and multimodal idiom understanding. Our submission focuses on the text-only track. Specifically, we employ an ensemble of three large language models (LLMs) to directly perform the presented image ranking task. Each model independently produces a ranking of the candidate images, and we aggregate their outputs using a hard voting strategy to determine the final prediction. This ensemble learning framework leverages the complementary strengths of different LLMs, improving robustness and reducing the variance of individual model predictions.
Anthology ID:
2026.mwe-1.27
Volume:
Proceedings of the 22nd Workshop on Multiword Expressions (MWE 2026)
Month:
March
Year:
2026
Address:
Rabat, Marocco
Editors:
Atul Kr. Ojha, Verginica Barbu Mititelu, Mathieu Constant, Ivelina Stoyanova, A. Seza Doğruöz, Alexandre Rademaker
Venues:
MWE | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
203–207
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.mwe-1.27/
DOI:
Bibkey:
Cite (ACL):
Le Qiu, Yu-Yin Hsu, and Emmanuele Chersoni. 2026. LST at MWE-2026 AdMIRe 2: Advancing Multimodal Idiomaticity Representation. In Proceedings of the 22nd Workshop on Multiword Expressions (MWE 2026), pages 203–207, Rabat, Marocco. Association for Computational Linguistics.
Cite (Informal):
LST at MWE-2026 AdMIRe 2: Advancing Multimodal Idiomaticity Representation (Qiu et al., MWE 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.mwe-1.27.pdf