@inproceedings{simon-jyothi-2025-deft,
title = "{D}e{FT}-{X}: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer",
author = "Simon, Sona Elza and
Jyothi, Preethi",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.100/",
doi = "10.18653/v1/2025.findings-emnlp.100",
pages = "1895--1909",
ISBN = "979-8-89176-335-7",
abstract = "Effective cross-lingual transfer remains a critical challenge in scaling the benefits of large language models from high-resource to low-resource languages. Towards this goal, prior studies have explored many approaches to combine task knowledge from task-specific data in a (high-resource) source language and language knowledge from unlabeled text in a (low-resource) target language. One notable approach proposed composable sparse fine-tuning (SFT) for cross-lingual transfer that learns task-specific and language-specific sparse masks to select a subset of the pretrained model{'}s parameters that are further fine-tuned. These sparse fine-tuned vectors (SFTs) are subsequently composed with the pretrained model to facilitate zero-shot cross-lingual transfer to a task in a target language, using only task-specific data from a source language. These sparse masks for SFTs were identified using a simple magnitude-based pruning. In our work, we introduce DeFT-X, a novel composable SFT approach that denoises the weight matrices of a pretrained model before magnitude pruning using singular value decomposition, thus yielding more robust SFTs. We evaluate DeFT-X on a diverse set of extremely low-resource languages for sentiment classification (NusaX) and natural language inference (AmericasNLI) and demonstrate that it performs at par or outperforms SFT and other prominent cross-lingual transfer baselines."
}Markdown (Informal)
[DeFT-X: Denoised Sparse Fine-Tuning for Zero-Shot Cross-Lingual Transfer](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.100/) (Simon & Jyothi, Findings 2025)
ACL