Instituto de Telecomunicações at IWSLT 2025: Aligning Small-Scale Speech and Language Models for Speech-to-Text Learning
Giuseppe Attanasio, Sonal Sannigrahi, Ben Peters, André Filipe Torres Martins
Abstract
This paper presents Instituto de Telecomunicações’s submission to the IWSLT 2025 Shared Task on Instruction Following Speech Processing. We submit results for the Short Track, i.e., speech recognition, translation, and spoken question answering. Our model is a unified speech-to-text model that integrates a pretrained continuous speech encoder and text decoder through a first phase of modality alignment and a second phase of instruction fine-tuning. Crucially, we focus on using small-scale language model backbones (< 2B) and restrict to high-quality, CC-BY data along with synthetic data generation to supplement existing resources.- Anthology ID:
- 2025.iwslt-1.36
- Volume:
- Proceedings of the 22nd International Conference on Spoken Language Translation (IWSLT 2025)
- Month:
- July
- Year:
- 2025
- Address:
- Vienna, Austria (in-person and online)
- Editors:
- Elizabeth Salesky, Marcello Federico, Antonis Anastasopoulos
- Venues:
- IWSLT | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 347–353
- Language:
- URL:
- https://preview.aclanthology.org/acl25-workshop-ingestion/2025.iwslt-1.36/
- DOI:
- Cite (ACL):
- Giuseppe Attanasio, Sonal Sannigrahi, Ben Peters, and André Filipe Torres Martins. 2025. Instituto de Telecomunicações at IWSLT 2025: Aligning Small-Scale Speech and Language Models for Speech-to-Text Learning. In Proceedings of the 22nd International Conference on Spoken Language Translation (IWSLT 2025), pages 347–353, Vienna, Austria (in-person and online). Association for Computational Linguistics.
- Cite (Informal):
- Instituto de Telecomunicações at IWSLT 2025: Aligning Small-Scale Speech and Language Models for Speech-to-Text Learning (Attanasio et al., IWSLT 2025)
- PDF:
- https://preview.aclanthology.org/acl25-workshop-ingestion/2025.iwslt-1.36.pdf