@inproceedings{xu-murray-2022-por,
title = "Por Qu{\'e} N{\~a}o Utiliser Alla Spr{\r{a}}k? Mixed Training with Gradient Optimization in Few-Shot Cross-Lingual Transfer",
author = "Xu, Haoran and
Murray, Kenton",
editor = "Carpuat, Marine and
de Marneffe, Marie-Catherine and
Meza Ruiz, Ivan Vladimir",
booktitle = "Findings of the Association for Computational Linguistics: NAACL 2022",
month = jul,
year = "2022",
address = "Seattle, United States",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2022.findings-naacl.157/",
doi = "10.18653/v1/2022.findings-naacl.157",
pages = "2043--2059",
abstract = "The current state-of-the-art for few-shot cross-lingual transfer learning first trains on abundant labeled data in the source language and then fine-tunes with a few examples on the target language, termed target-adapting. Though this has been demonstrated to work on a variety of tasks, in this paper we show some deficiencies of this approach and propose a one-step mixed training method that trains on both source and target data with stochastic gradient surgery, a novel gradient-level optimization. Unlike the previous studies that focus on one language at a time when target-adapting, we use one model to handle all target languages simultaneously to avoid excessively language-specific models. Moreover, we discuss the unreality of utilizing large target development sets for model selection in previous literature. We further show that our method is both development-free for target languages, and is also able to escape from overfitting issues. We conduct a large-scale experiment on 4 diverse NLP tasks across up to 48 languages. Our proposed method achieves state-of-the-art performance on all tasks and outperforms target-adapting by a large margin, especially for languages that are linguistically distant from the source language, e.g., 7.36{\%} F1 absolute gain on average for the NER task, up to 17.60{\%} on Punjabi."
}
Markdown (Informal)
[Por Qué Não Utiliser Alla Språk? Mixed Training with Gradient Optimization in Few-Shot Cross-Lingual Transfer](https://preview.aclanthology.org/landing_page/2022.findings-naacl.157/) (Xu & Murray, Findings 2022)
ACL