Merge and Conquer: Instructing Multilingual Models by Adding Target Language Weights
Eneko Valero, Maria Ribalta i Albado, Oscar Sainz, Naiara Perez, German Rigau
Abstract
Large Language Models (LLMs) remain heavily centered on English, with limited performance in low-resource languages. Existing adaptation approaches, such as continual pre-training, demand significant computational resources. In the case of instructed models, high-quality instruction data is also required, both of which are often inaccessible for low-resource language communities. Under these constraints, model merging offers a lightweight alternative, but its potential in low-resource contexts has not been systematically explored. In this work, we explore whether it is possible to transfer language knowledge to an instruction-tuned LLM by merging it with a language-specific base model, thereby eliminating the need of language-specific instructions and repeated fine-tuning processes whenever stronger instructed variants become available. Through experiments covering four Iberian languages (Basque, Catalan, Galician, and Spanish) and two model families, we show that merging enables effective instruction-following behavior in new languages and even supports multilingual capability through the combination of multiple language-specific models. Our results indicate that model merging is a viable and efficient alternative to traditional adaptation methods for low-resource languages, achieving competitive performance while greatly reducing computational cost.- Anthology ID:
- 2026.lrec-main.800
- Volume:
- Proceedings of the Fifteenth Language Resources and Evaluation Conference
- Month:
- May
- Year:
- 2026
- Address:
- Palma de Mallorca, Spain
- Editors:
- Stelios Piperidis, Núria Bel, Henk van den Heuvel, Nancy Ide, Simon Krek, Antonio Toral
- Venue:
- LREC
- SIG:
- Publisher:
- ELRA Language Resource Association
- Note:
- Pages:
- 10192–10207
- Language:
- URL:
- https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.800/
- DOI:
- Cite (ACL):
- Eneko Valero, Maria Ribalta i Albado, Oscar Sainz, Naiara Perez, and German Rigau. 2026. Merge and Conquer: Instructing Multilingual Models by Adding Target Language Weights. International Conference on Language Resources and Evaluation, main:10192–10207.
- Cite (Informal):
- Merge and Conquer: Instructing Multilingual Models by Adding Target Language Weights (Valero et al., LREC 2026)
- PDF:
- https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.800.pdf