Miguel Claramunt
2025
From SALAMANDRA to SALAMANDRATA: BSC Submission for WMT25 General Machine Translation Shared Task
Javier Garcia Gilabert
|
Xixian Liao
|
Severino Da Dalt
|
Ella Bohman
|
Audrey Mash
|
Francesca De Luca Fornaciari
|
Irene Baucells
|
Joan Llop
|
Miguel Claramunt
|
Carlos Escolano
|
Maite Melero
Proceedings of the Tenth Conference on Machine Translation
In this paper, we present the SalamandraTA family of models, an improved iteration of Salamandra LLMs (Gonzalez-Agirre et al., 2025) specifically trained to achieve strong performance in translation-related tasks for 38 European languages. SalamandraTA comes in two scales: 2B and 7B parameters. For both versions, we applied the same training recipe with a first step of continual pre-training on parallel data, and a second step of supervised fine-tuning on high-quality instructions.The BSC submission to the WMT25 General Machine Translation shared task is based on the 7B variant of SalamandraTA. We first extended the model vocabulary to support the additional non-European languages included in the task. This was followed by a second phase of continual pretraining and supervised fine-tuning, carefully designed to optimize performance across all translation directions for this year’s shared task. For decoding, we employed two quality-aware strategies: Minimum Bayes Risk Decoding and Translation Reranking using Comet and Comet-kiwi.We publicly release both the 2B and 7B versions of SalamandraTA, along with the newer SalamandraTA-v2 model, on Hugging Face.
Search
Fix author
Co-authors
- Irene Baucells 1
- Ella Bohman 1
- Severino Da Dalt 1
- Carlos Escolano 1
- Francesca De Luca Fornaciari 1
- show all...
Venues
- wmt1