@inproceedings{buma-etal-2025-two,
title = "Two Step Automatic Post Editing of Patent Machine Translation based on Pre-trained Encoder Models and {LLM}s",
author = "Buma, Kosei and
Utsuro, Takehito and
Nagata, Masaaki",
editor = "T.y.s.s, Santosh and
Shimizu, Shuichiro and
Gong, Yifan",
booktitle = "The 14th International Joint Conference on Natural Language Processing and The 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics",
month = dec,
year = "2025",
address = "Mumbai, India",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.ijcnlp-srw.19/",
pages = "218--231",
ISBN = "979-8-89176-304-3",
abstract = "We study automatic post-editing for patent translation, where accuracy and traceability are critical, and propose a two-step pipeline that combines a multilingual encoder for token-level error detection with an LLM for targeted correction. As no word-level annotations exist for Japanese{--}English patents, we create supervised data by injecting synthetic errors into parallel patent sentences and fine-tune mBERT, XLM-RoBERTa, and mDeBERTa as detectors. In the second stage, GPT-4o is prompted to revise translations either freely or under a restricted policy that allows edits only on detector-marked spans. For error detection, evaluation on synthetic errors shows that encoder-based detectors outperform LLMs in both F1 and MCC. For error correction, tests on synthetic, repetition, and omission datasets demonstrate statistically significant BLEU gains over LLM methods for synthetic and repetition errors, while omission errors remain challenging. Overall, pairing compact encoders with an LLM enables more accurate and controllable post-editing for key patent error types, reducing unnecessary rewrites via restricted edits. Future work will focus on strengthening omission modeling to better detect and correct missing content."
}Markdown (Informal)
[Two Step Automatic Post Editing of Patent Machine Translation based on Pre-trained Encoder Models and LLMs](https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.ijcnlp-srw.19/) (Buma et al., IJCNLP 2025)
ACL