@inproceedings{geigle-etal-2021-tuda,
title = "{TUD}a at {WMT}21: Sentence-Level Direct Assessment with Adapters",
author = {Geigle, Gregor and
Stadtm{\"u}ller, Jonas and
Zhao, Wei and
Pfeiffer, Jonas and
Eger, Steffen},
booktitle = "Proceedings of the Sixth Conference on Machine Translation",
month = nov,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.wmt-1.95",
pages = "911--919",
abstract = "This paper presents our submissions to the WMT2021 Shared Task on Quality Estimation, Task 1 \textit{Sentence-Level Direct Assessment}. While top-performing approaches utilize massively multilingual Transformer-based language models which have been pre-trained on all target languages of the task, the resulting insights are limited, as it is unclear how well the approach performs on languages unseen during pre-training; more problematically, these approaches do not provide any solutions for \textit{extending} the model to new languages or unseen scripts{---}arguably one of the objectives of this shared task. In this work, we thus focus on utilizing massively multilingual language models which only \textit{partly} cover the target languages during their pre-training phase. We extend the model to new languages and unseen scripts using recent adapter-based methods and achieve on par performance or even surpass models pre-trained on the respective languages.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="geigle-etal-2021-tuda">
<titleInfo>
<title>TUDa at WMT21: Sentence-Level Direct Assessment with Adapters</title>
</titleInfo>
<name type="personal">
<namePart type="given">Gregor</namePart>
<namePart type="family">Geigle</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jonas</namePart>
<namePart type="family">Stadtmüller</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Wei</namePart>
<namePart type="family">Zhao</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Jonas</namePart>
<namePart type="family">Pfeiffer</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Steffen</namePart>
<namePart type="family">Eger</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-nov</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Sixth Conference on Machine Translation</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>This paper presents our submissions to the WMT2021 Shared Task on Quality Estimation, Task 1 Sentence-Level Direct Assessment. While top-performing approaches utilize massively multilingual Transformer-based language models which have been pre-trained on all target languages of the task, the resulting insights are limited, as it is unclear how well the approach performs on languages unseen during pre-training; more problematically, these approaches do not provide any solutions for extending the model to new languages or unseen scripts—arguably one of the objectives of this shared task. In this work, we thus focus on utilizing massively multilingual language models which only partly cover the target languages during their pre-training phase. We extend the model to new languages and unseen scripts using recent adapter-based methods and achieve on par performance or even surpass models pre-trained on the respective languages.</abstract>
<identifier type="citekey">geigle-etal-2021-tuda</identifier>
<location>
<url>https://aclanthology.org/2021.wmt-1.95</url>
</location>
<part>
<date>2021-nov</date>
<extent unit="page">
<start>911</start>
<end>919</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T TUDa at WMT21: Sentence-Level Direct Assessment with Adapters
%A Geigle, Gregor
%A Stadtmüller, Jonas
%A Zhao, Wei
%A Pfeiffer, Jonas
%A Eger, Steffen
%S Proceedings of the Sixth Conference on Machine Translation
%D 2021
%8 nov
%I Association for Computational Linguistics
%C Online
%F geigle-etal-2021-tuda
%X This paper presents our submissions to the WMT2021 Shared Task on Quality Estimation, Task 1 Sentence-Level Direct Assessment. While top-performing approaches utilize massively multilingual Transformer-based language models which have been pre-trained on all target languages of the task, the resulting insights are limited, as it is unclear how well the approach performs on languages unseen during pre-training; more problematically, these approaches do not provide any solutions for extending the model to new languages or unseen scripts—arguably one of the objectives of this shared task. In this work, we thus focus on utilizing massively multilingual language models which only partly cover the target languages during their pre-training phase. We extend the model to new languages and unseen scripts using recent adapter-based methods and achieve on par performance or even surpass models pre-trained on the respective languages.
%U https://aclanthology.org/2021.wmt-1.95
%P 911-919
Markdown (Informal)
[TUDa at WMT21: Sentence-Level Direct Assessment with Adapters](https://aclanthology.org/2021.wmt-1.95) (Geigle et al., WMT 2021)
ACL