NHK Submission to WAT 2025: Leveraging Preference Optimization for Article-level Japanese–English News Translation

Hideya Mino, Rei Endo, Yoshihiko Kawai


Abstract
This paper describes our submission to the Japanese English Article-level News Translation Task at WAT 2025. In this task, participants were provided with a small but high-quality parallel corpus along with two intermediate English translations: a literal translation and a style-adapted translation. To effectively exploit these limited training data, our system employs a large language model (LLM) trained via supervised fine-tuning (SFT) followed by Direct Preference Optimization (DPO) that is a preference learning technique for aligning model outputs with professional-quality references. By leveraging literal and style-adapted intermediate translations as negative (rejected) samples and human-edited English articles as positive (chosen) samples in DPO training, we achieved notable improvements in translation quality. We evaluate our approach using BLEU scores and human assessments.
Anthology ID:
2025.wat-1.9
Volume:
Proceedings of the Twelfth Workshop on Asian Translation (WAT 2025)
Month:
December
Year:
2025
Address:
Mumbai, India
Editors:
Toshiaki Nakazawa, Isao Goto
Venues:
WAT | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
98–102
Language:
URL:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.wat-1.9/
DOI:
Bibkey:
Cite (ACL):
Hideya Mino, Rei Endo, and Yoshihiko Kawai. 2025. NHK Submission to WAT 2025: Leveraging Preference Optimization for Article-level Japanese–English News Translation. In Proceedings of the Twelfth Workshop on Asian Translation (WAT 2025), pages 98–102, Mumbai, India. Association for Computational Linguistics.
Cite (Informal):
NHK Submission to WAT 2025: Leveraging Preference Optimization for Article-level Japanese–English News Translation (Mino et al., WAT 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.wat-1.9.pdf