@inproceedings{cui-etal-2025-quality,
title = "Quality-aware Neural Machine Translation with Self-evaluation",
author = "Cui, Jiajia and
Mu, Lingling and
Liu, Qiuhui and
Xu, Hongfei",
editor = "Sun, Maosong and
Duan, Peiyong and
Liu, Zhiyuan and
Xu, Ruifeng and
Sun, Weiwei",
booktitle = "Proceedings of the 24th {C}hina National Conference on Computational Linguistics ({CCL} 2025)",
month = aug,
year = "2025",
address = "Jinan, China",
publisher = "Chinese Information Processing Society of China",
url = "https://preview.aclanthology.org/ingest-ccl/2025.ccl-1.87/",
pages = "1178--1187",
abstract = "``The performance of neural machine translation relies on a large amount of data, but crawled sentence pairs are of different quality. The low-quality sentence pairs may provide helpful translation knowledge but also teach the model to generate low-quality translations. Making the model aware of the quality of training instances may help the model distinguish between good and bad translations while leveraging the translation knowledge. In this paper, we evaluate the quality of training instances with the average per-token loss (negative log-likelihood) from translation mod-els, convert the quality scores into embeddings through vector interpolation and feed the quality embedding into the translation model during its training. We ask the model to decode with the best quality score to generate good translations during inference. Experiments on the IWSLT 14 German to English, WMT 14 English to German and WMT 22 English to Japanese translation tasks show that our method can effectively lead to consistent and significant improvements across multiple metrics.''"
}Markdown (Informal)
[Quality-aware Neural Machine Translation with Self-evaluation](https://preview.aclanthology.org/ingest-ccl/2025.ccl-1.87/) (Cui et al., CCL 2025)
ACL