@inproceedings{dai-etal-2022-bertology,
title = "{BERT}ology for Machine Translation: What {BERT} Knows about Linguistic Difficulties for Translation",
author = "Dai, Yuqian and
de Kamps, Marc and
Sharoff, Serge",
editor = "Calzolari, Nicoletta and
B{\'e}chet, Fr{\'e}d{\'e}ric and
Blache, Philippe and
Choukri, Khalid and
Cieri, Christopher and
Declerck, Thierry and
Goggi, Sara and
Isahara, Hitoshi and
Maegaard, Bente and
Mariani, Joseph and
Mazo, H{\'e}l{\`e}ne and
Odijk, Jan and
Piperidis, Stelios",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2022.lrec-1.719/",
pages = "6674--6690",
abstract = "Pre-trained transformer-based models, such as BERT, have shown excellent performance in most natural language processing benchmark tests, but we still lack a good understanding of the linguistic knowledge of BERT in Neural Machine Translation (NMT). Our work uses syntactic probes and Quality Estimation (QE) models to analyze the performance of BERT`s syntactic dependencies and their impact on machine translation quality, exploring what kind of syntactic dependencies are difficult for NMT engines based on BERT. While our probing experiments confirm that pre-trained BERT {\textquotedblleft}knows{\textquotedblright} about syntactic dependencies, its ability to recognize them often decreases after fine-tuning for NMT tasks. We also detect a relationship between syntactic dependencies in three languages and the quality of their translations, which shows which specific syntactic dependencies are likely to be a significant cause of low-quality translations."
}
Markdown (Informal)
[BERTology for Machine Translation: What BERT Knows about Linguistic Difficulties for Translation](https://preview.aclanthology.org/add-emnlp-2024-awards/2022.lrec-1.719/) (Dai et al., LREC 2022)
ACL