Zhengyang Liu
2025
PRIM: Towards Practical In-Image Multilingual Machine Translation
Yanzhi Tian
|
Zeming Liu
|
Zhengyang Liu
|
Chong Feng
|
Xin Li
|
Heyan Huang
|
Yuhang Guo
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
In-Image Machine Translation (IIMT) aims to translate images containing texts from one language to another. Current research of end-to-end IIMT mainly conducts on synthetic data, with simple background, single font, fixed text position, and bilingual translation, which can not fully reflect real world, causing a significant gap between the research and practical conditions. To facilitate research of IIMT in real-world scenarios, we explore Practical In-Image Multilingual Machine Translation (IIMMT). In order to convince the lack of publicly available data, we annotate the PRIM dataset, which contains real-world captured one-line text images with complex background, various fonts, diverse text positions, and supports multilingual translation directions. We propose an end-to-end model VisTrans to handle the challenge of practical conditions in PRIM, which processes visual text and background information in the image separately, ensuring the capability of multilingual translation while improving the visual quality. Experimental results indicate the VisTrans achieves a better translation quality and visual effect compared to other models. The code and dataset are available at: https://github.com/BITHLP/PRIM.
Exploring In-Image Machine Translation with Real-World Background
Yanzhi Tian
|
Zeming Liu
|
Zhengyang Liu
|
Yuhang Guo
Findings of the Association for Computational Linguistics: ACL 2025
In-Image Machine Translation (IIMT) aims to translate texts within images from one language to another. Previous research on IIMT was primarily conducted on simplified scenarios such as images of one-line text with black font in white backgrounds, which is far from reality and impractical for applications in the real world. To make IIMT research practically valuable, it is essential to consider a complex scenario where the text backgrounds are derived from real-world images. To facilitate research of complex scenarios IIMT, we design an IIMT dataset that includes subtitle text with a real-world background. However, previous IIMT models perform inadequately in complex scenarios. To address the issue, we propose the DebackX model, which separates the background and text-image from the source image, performs translation on the text-image directly, and fuses the translated text-image with the background to generate the target image. Experimental results show that our model achieves improvements in both translation quality and visual effect.
Search
Fix author
Co-authors
- Yuhang Guo (郭宇航) 2
- Zeming Liu 2
- Yanzhi Tian 2
- Chong Feng (冯冲) 1
- He-Yan Huang 1
- show all...
- Xin Li 1