The Effects of Language Token Prefixing for Multilingual Machine Translation

Rachel Wicks, Kevin Duh


Abstract
Machine translation traditionally refers to translating from a single source language into a single target language. In recent years, the field has moved towards large neural models either translating from or into many languages. The model must be correctly cued to translate into the correct target language.This is typically done by prefixing language tokens onto the source or target sequence. The location and content of the prefix can vary and many use different approaches without much justification towards one approach or another. As a guidance to future researchers and directions for future work, we present a series of experiments that show how the positioning and type of a target language prefix token effects translation performance. We show that source side prefixes improve performance. Further, we find that the best language information to denote via tokens depends on the supported language set.
Anthology ID:
2022.aacl-short.19
Volume:
Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)
Month:
November
Year:
2022
Address:
Online only
Venues:
AACL | IJCNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
148–153
Language:
URL:
https://aclanthology.org/2022.aacl-short.19
DOI:
Bibkey:
Cite (ACL):
Rachel Wicks and Kevin Duh. 2022. The Effects of Language Token Prefixing for Multilingual Machine Translation. In Proceedings of the 2nd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics and the 12th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 148–153, Online only. Association for Computational Linguistics.
Cite (Informal):
The Effects of Language Token Prefixing for Multilingual Machine Translation (Wicks & Duh, AACL-IJCNLP 2022)
Copy Citation:
PDF:
https://preview.aclanthology.org/auto-file-uploads/2022.aacl-short.19.pdf