@inproceedings{wu-etal-2025-english,
title = "From {E}nglish to Second Language Mastery: Enhancing {LLM}s with Cross-Lingual Continued Instruction Tuning",
author = "Wu, Linjuan and
Wei, Hao-Ran and
Yang, Baosong and
Lu, Weiming",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1121/",
pages = "23006--23023",
ISBN = "979-8-89176-251-0",
abstract = "Supervised Fine-Tuning (SFT) with translated instruction data effectively adapts Large Language Models (LLMs) from English to non-English languages. We introduce Cross-Lingual Continued Instruction Tuning (X-CIT), which fully leverages translation-based parallel instruction data to enhance cross-lingual adaptability. X-CIT emulates the human process of second language acquisition and is guided by Chomsky{'}s Principles and Parameters Theory. It first fine-tunes the LLM on English instruction data to establish foundational capabilities (i.e. Principles), then continues with target language translation and customized chat-instruction data to adjust ``parameters'' specific to the target language. This chat-instruction data captures alignment information in translated parallel data, guiding the model to initially think and respond in its native language before transitioning to the target language. To further mimic human learning progression, we incorporate Self-Paced Learning (SPL) during continued training, allowing the model to advance from simple to complex tasks. Implemented on Llama-2-7B across five languages, X-CIT was evaluated against three objective benchmarks and an LLM-as-a-judge benchmark, improving the strongest baseline by an average of 1.97{\%} and 8.2{\%} in these two benchmarks, respectively."
}
Markdown (Informal)
[From English to Second Language Mastery: Enhancing LLMs with Cross-Lingual Continued Instruction Tuning](https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1121/) (Wu et al., ACL 2025)
ACL