A Comparative Analysis of Unsupervised Language Adaptation Methods

Gil Rocha, Henrique Lopes Cardoso


Abstract
To overcome the lack of annotated resources in less-resourced languages, recent approaches have been proposed to perform unsupervised language adaptation. In this paper, we explore three recent proposals: Adversarial Training, Sentence Encoder Alignment and Shared-Private Architecture. We highlight the differences of these approaches in terms of unlabeled data requirements and capability to overcome additional domain shift in the data. A comparative analysis in two different tasks is conducted, namely on Sentiment Classification and Natural Language Inference. We show that adversarial training methods are more suitable when the source and target language datasets contain other variations in content besides the language shift. Otherwise, sentence encoder alignment methods are very effective and can yield scores on the target language that are close to the source language scores.
Anthology ID:
D19-6102
Volume:
Proceedings of the 2nd Workshop on Deep Learning Approaches for Low-Resource NLP (DeepLo 2019)
Month:
November
Year:
2019
Address:
Hong Kong, China
Editors:
Colin Cherry, Greg Durrett, George Foster, Reza Haffari, Shahram Khadivi, Nanyun Peng, Xiang Ren, Swabha Swayamdipta
Venue:
WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
11–21
Language:
URL:
https://aclanthology.org/D19-6102
DOI:
10.18653/v1/D19-6102
Bibkey:
Cite (ACL):
Gil Rocha and Henrique Lopes Cardoso. 2019. A Comparative Analysis of Unsupervised Language Adaptation Methods. In Proceedings of the 2nd Workshop on Deep Learning Approaches for Low-Resource NLP (DeepLo 2019), pages 11–21, Hong Kong, China. Association for Computational Linguistics.
Cite (Informal):
A Comparative Analysis of Unsupervised Language Adaptation Methods (Rocha & Lopes Cardoso, 2019)
Copy Citation:
PDF:
https://preview.aclanthology.org/emnlp-22-attachments/D19-6102.pdf