ChuenSumi at SemEval-2025 Task 1: Sentence Transformer Models and Processing Idiomacity

Sumiko Teng, Chuen Shin Yong


Abstract
This paper participates Task 1 of SemEval2025, specifically Subtask A’s English Text-Only track, where we develop a model to rank text descriptions of images with respect to how well it represents a the use of a given multi-word expression in its respective context sentence. We trained sentence transformer models from huggingface to rank the text descriptions, finding the RoBERTa model to be the better performing model. For the final evaluation, the fine-tuned RoBERTa model achieved an accuracy of 0.4 for the first developer’s evaluation set, and 0.2 for the second, ranking 9th in the English Text Only category for Subtask A. Overall, our results show that a vanilla sentence transformerapproach performs adequately in the task and processing idioms. They also suggest that RoBERTa models may be stronger in idiom processing than other models.
Anthology ID:
2025.semeval-1.18
Volume:
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Sara Rosenthal, Aiala Rosá, Debanjan Ghosh, Marcos Zampieri
Venues:
SemEval | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
122–126
Language:
URL:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.18/
DOI:
Bibkey:
Cite (ACL):
Sumiko Teng and Chuen Shin Yong. 2025. ChuenSumi at SemEval-2025 Task 1: Sentence Transformer Models and Processing Idiomacity. In Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025), pages 122–126, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
ChuenSumi at SemEval-2025 Task 1: Sentence Transformer Models and Processing Idiomacity (Teng & Yong, SemEval 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.18.pdf