@inproceedings{wang-etal-2021-tencent-ai,
    title = "Tencent {AI} Lab Machine Translation Systems for the {WMT}21 Biomedical Translation Task",
    author = "Wang, Xing  and
      Tu, Zhaopeng  and
      Shi, Shuming",
    editor = "Barrault, Loic  and
      Bojar, Ondrej  and
      Bougares, Fethi  and
      Chatterjee, Rajen  and
      Costa-jussa, Marta R.  and
      Federmann, Christian  and
      Fishel, Mark  and
      Fraser, Alexander  and
      Freitag, Markus  and
      Graham, Yvette  and
      Grundkiewicz, Roman  and
      Guzman, Paco  and
      Haddow, Barry  and
      Huck, Matthias  and
      Yepes, Antonio Jimeno  and
      Koehn, Philipp  and
      Kocmi, Tom  and
      Martins, Andre  and
      Morishita, Makoto  and
      Monz, Christof",
    booktitle = "Proceedings of the Sixth Conference on Machine Translation",
    month = nov,
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2021.wmt-1.89/",
    pages = "874--878",
    abstract = "This paper describes the Tencent AI Lab submission of the WMT2021 shared task on biomedical translation in eight language directions: English-German, English-French, English-Spanish and English-Russian. We utilized different Transformer architectures, pretraining and back-translation strategies to improve translation quality. Concretely, we explore mBART (Liu et al., 2020) to demonstrate the effectiveness of the pretraining strategy. Our submissions (Tencent AI Lab Machine Translation, TMT) in German/French/Spanish{\ensuremath{\Rightarrow}}English are ranked 1st respectively according to the official evaluation results in terms of BLEU scores."
}Markdown (Informal)
[Tencent AI Lab Machine Translation Systems for the WMT21 Biomedical Translation Task](https://preview.aclanthology.org/ingest-emnlp/2021.wmt-1.89/) (Wang et al., WMT 2021)
ACL