@inproceedings{kondo-etal-2021-machine,
title = "Machine Translation with Pre-specified Target-side Words Using a Semi-autoregressive Model",
author = "Kondo, Seiichiro and
Koyama, Aomi and
Kiyuna, Tomoshige and
Hirasawa, Tosho and
Komachi, Mamoru",
booktitle = "Proceedings of the 8th Workshop on Asian Translation (WAT2021)",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.wat-1.5",
doi = "10.18653/v1/2021.wat-1.5",
pages = "68--73",
abstract = "We introduce our TMU Japanese-to-English system, which employs a semi-autoregressive model, to tackle the WAT 2021 restricted translation task. In this task, we translate an input sentence with the constraint that some words, called restricted target vocabularies (RTVs), must be contained in the output sentence. To satisfy this constraint, we use a semi-autoregressive model, namely, RecoverSAT, due to its ability (known as {``}forced translation{''}) to insert specified words into the output sentence. When using {``}forced translation,{''} the order of inserting RTVs is a critical problem. In this work, we aligned the source sentence and the corresponding RTVs using GIZA++. In our system, we obtain word alignment between a source sentence and the corresponding RTVs and then sort the RTVs in the order of their corresponding words or phrases in the source sentence. Using the model with sorted order RTVs, we succeeded in inserting all the RTVs into output sentences in more than 96{\%} of the test sentences. Moreover, we confirmed that sorting RTVs improved the BLEU score compared with random order RTVs.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="kondo-etal-2021-machine">
<titleInfo>
<title>Machine Translation with Pre-specified Target-side Words Using a Semi-autoregressive Model</title>
</titleInfo>
<name type="personal">
<namePart type="given">Seiichiro</namePart>
<namePart type="family">Kondo</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Aomi</namePart>
<namePart type="family">Koyama</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Tomoshige</namePart>
<namePart type="family">Kiyuna</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Tosho</namePart>
<namePart type="family">Hirasawa</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mamoru</namePart>
<namePart type="family">Komachi</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2021-aug</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 8th Workshop on Asian Translation (WAT2021)</title>
</titleInfo>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Online</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>We introduce our TMU Japanese-to-English system, which employs a semi-autoregressive model, to tackle the WAT 2021 restricted translation task. In this task, we translate an input sentence with the constraint that some words, called restricted target vocabularies (RTVs), must be contained in the output sentence. To satisfy this constraint, we use a semi-autoregressive model, namely, RecoverSAT, due to its ability (known as “forced translation”) to insert specified words into the output sentence. When using “forced translation,” the order of inserting RTVs is a critical problem. In this work, we aligned the source sentence and the corresponding RTVs using GIZA++. In our system, we obtain word alignment between a source sentence and the corresponding RTVs and then sort the RTVs in the order of their corresponding words or phrases in the source sentence. Using the model with sorted order RTVs, we succeeded in inserting all the RTVs into output sentences in more than 96% of the test sentences. Moreover, we confirmed that sorting RTVs improved the BLEU score compared with random order RTVs.</abstract>
<identifier type="citekey">kondo-etal-2021-machine</identifier>
<identifier type="doi">10.18653/v1/2021.wat-1.5</identifier>
<location>
<url>https://aclanthology.org/2021.wat-1.5</url>
</location>
<part>
<date>2021-aug</date>
<extent unit="page">
<start>68</start>
<end>73</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Machine Translation with Pre-specified Target-side Words Using a Semi-autoregressive Model
%A Kondo, Seiichiro
%A Koyama, Aomi
%A Kiyuna, Tomoshige
%A Hirasawa, Tosho
%A Komachi, Mamoru
%S Proceedings of the 8th Workshop on Asian Translation (WAT2021)
%D 2021
%8 aug
%I Association for Computational Linguistics
%C Online
%F kondo-etal-2021-machine
%X We introduce our TMU Japanese-to-English system, which employs a semi-autoregressive model, to tackle the WAT 2021 restricted translation task. In this task, we translate an input sentence with the constraint that some words, called restricted target vocabularies (RTVs), must be contained in the output sentence. To satisfy this constraint, we use a semi-autoregressive model, namely, RecoverSAT, due to its ability (known as “forced translation”) to insert specified words into the output sentence. When using “forced translation,” the order of inserting RTVs is a critical problem. In this work, we aligned the source sentence and the corresponding RTVs using GIZA++. In our system, we obtain word alignment between a source sentence and the corresponding RTVs and then sort the RTVs in the order of their corresponding words or phrases in the source sentence. Using the model with sorted order RTVs, we succeeded in inserting all the RTVs into output sentences in more than 96% of the test sentences. Moreover, we confirmed that sorting RTVs improved the BLEU score compared with random order RTVs.
%R 10.18653/v1/2021.wat-1.5
%U https://aclanthology.org/2021.wat-1.5
%U https://doi.org/10.18653/v1/2021.wat-1.5
%P 68-73
Markdown (Informal)
[Machine Translation with Pre-specified Target-side Words Using a Semi-autoregressive Model](https://aclanthology.org/2021.wat-1.5) (Kondo et al., WAT 2021)
ACL