@inproceedings{wu-etal-2023-oolong,
title = "Oolong: Investigating What Makes Transfer Learning Hard with Controlled Studies",
author = "Wu, Zhengxuan and
Tamkin, Alex and
Papadimitriou, Isabel",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.198/",
doi = "10.18653/v1/2023.emnlp-main.198",
pages = "3280--3289",
abstract = "When we transfer a pretrained language model to a new language, there are many axes of variation that change at once. To disentangle the impact of different factors like syntactic similarity and vocabulary similarity, we propose a set of \textit{controlled transfer studies}: we systematically transform the language of the GLUE benchmark, altering one axis of crosslingual variation at a time, and then measure the resulting drops in a pretrained model{'}s downstream performance. We find that models can largely recover from syntactic-style shifts, but cannot recover from vocabulary misalignment and embedding matrix re-initialization, even with continued pretraining on 15 million tokens. Moreover, good-quality tokenizers in the transfer language do not make vocabulary alignment easier. Our experiments provide insights into the factors of cross-lingual transfer that researchers should most focus on when designing language transfer scenarios."
}
Markdown (Informal)
[Oolong: Investigating What Makes Transfer Learning Hard with Controlled Studies](https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.198/) (Wu et al., EMNLP 2023)
ACL