@inproceedings{gupta-2022-malm,
title = "{MALM}: Mixing Augmented Language Modeling for Zero-Shot Machine Translation",
author = "Gupta, Kshitij",
editor = {H{\"a}m{\"a}l{\"a}inen, Mika and
Alnajjar, Khalid and
Partanen, Niko and
Rueter, Jack},
booktitle = "Proceedings of the 2nd International Workshop on Natural Language Processing for Digital Humanities",
month = nov,
year = "2022",
address = "Taipei, Taiwan",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2022.nlp4dh-1.8/",
doi = "10.18653/v1/2022.nlp4dh-1.8",
pages = "53--58",
abstract = "Large pre-trained language models have brought remarkable progress in NLP. Pre-training and Fine-tuning have given state-of-art performance across tasks in text processing. Data Augmentation techniques have also helped build state-of-art models on low or zero resource tasks. Many works in the past have attempted at learning a single massively multilingual machine translation model for zero-shot translation. Although those translation models are producing correct translations, the main challenge is those models are producing the wrong languages for zero-shot translation. This work and its results indicate that prompt conditioned large models do not suffer from off-target language errors i.e. errors arising due to translation to wrong languages. We empirically demonstrate the effectiveness of self-supervised pre-training and data augmentation for zero-shot multi-lingual machine translation."
}
Markdown (Informal)
[MALM: Mixing Augmented Language Modeling for Zero-Shot Machine Translation](https://preview.aclanthology.org/jlcl-multiple-ingestion/2022.nlp4dh-1.8/) (Gupta, NLP4DH 2022)
ACL