CTYUN-AI at SemEval-2025 Task 1: Learning to Rank for Idiomatic Expressions

Yuming Fan, Dongming Yang, Zefeng Cai, Binghuai Lin


Abstract
We propose a multimodal framework integrating textual context and image caption analysis via systematic data augmentation and parameter-efficient fine-tuning. Our approach features: (1) option shuffling to eliminate positional bias, (2) lexical augmentation through synonym replacement and back-translation, and (3) optimized cross-modal ranking adaptation. The system ranks first in Portuguese (Top-1 Acc: 0.92) and second in English (Top-1 Acc: 0.87) on CodaBench. Experiments across 7B-72B models reveal 32B architectures achieve optimal capacity-trainability balance, while larger 72B models suffer from overfitting. Results demonstrate the limitations of GPT-4 knowledge distillation and emphasize controlled data augmentation for idiomatic language learning, advancing multimodal figurative language processing techniques.
Anthology ID:
2025.semeval-1.3
Volume:
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Sara Rosenthal, Aiala Rosá, Debanjan Ghosh, Marcos Zampieri
Venues:
SemEval | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
16–19
Language:
URL:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.3/
DOI:
Bibkey:
Cite (ACL):
Yuming Fan, Dongming Yang, Zefeng Cai, and Binghuai Lin. 2025. CTYUN-AI at SemEval-2025 Task 1: Learning to Rank for Idiomatic Expressions. In Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025), pages 16–19, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
CTYUN-AI at SemEval-2025 Task 1: Learning to Rank for Idiomatic Expressions (Fan et al., SemEval 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.3.pdf