@inproceedings{lee-etal-2025-semantic-aware,
title = "Semantic Aware Linear Transfer by Recycling Pre-trained Language Models for Cross-lingual Transfer",
author = "Lee, Seungyoon and
Hong, Seongtae and
Moon, Hyeonseok and
Lim, Heuiseok",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.832/",
doi = "10.18653/v1/2025.findings-acl.832",
pages = "16180--16193",
ISBN = "979-8-89176-256-5",
abstract = "Large Language Models (LLMs) are increasingly incorporating multilingual capabilities, fueling the demand to transfer them into target language-specific models. However, most approaches, which blend the source model{'}s embedding by replacing the source vocabulary with the target language-specific vocabulary, may constrain expressive capacity in the target language since the source model is predominantly trained on English data. In this paper, we propose Semantic Aware Linear Transfer (SALT), a novel cross-lingual transfer technique that recycles embeddings from target language Pre-trained Language Models (PLMs) to transmit the deep representational strengths of PLM-derived embedding to LLMs. SALT derives unique regression lines based on the similarity in the overlap of the source and target vocabularies to handle each non-overlapping token{'}s embedding space. Our extensive experiments show that SALT significantly outperforms other transfer methods, achieving lower loss and faster convergence during language adaptation. Notably, SALT achieves remarkable performance in cross-lingual understanding setups compared to other methods. Furthermore, we highlight the scalable use of PLMs to enhance the functionality of contemporary LLMs by conducting experiments with varying architectures."
}
Markdown (Informal)
[Semantic Aware Linear Transfer by Recycling Pre-trained Language Models for Cross-lingual Transfer](https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.832/) (Lee et al., Findings 2025)
ACL