2024
pdf
Exploring the Necessity of Visual Modality in Multimodal Machine Translation using Authentic Datasets
Zi Long
|
ZhenHao Tang
|
Xianghua Fu
|
Jian Chen
|
Shilong Hou
|
Jinze Lyu
Proceedings of the 17th Workshop on Building and Using Comparable Corpora (BUCC) @ LREC-COLING 2024
2022
pdf
abs
Multimodal Neural Machine Translation with Search Engine Based Image Retrieval
ZhenHao Tang
|
XiaoBing Zhang
|
Zi Long
|
XiangHua Fu
Proceedings of the 9th Workshop on Asian Translation
Recently, numbers of works shows that the performance of neural machine translation (NMT) can be improved to a certain extent with using visual information. However, most of these conclusions are drawn from the analysis of experimental results based on a limited set of bilingual sentence-image pairs, such as Multi30K.In these kinds of datasets, the content of one bilingual parallel sentence pair must be well represented by a manually annotated image,which is different with the actual translation situation. we propose an open-vocabulary image retrieval methods to collect descriptive images for bilingual parallel corpus using image search engine, and we propose text-aware attentive visual encoder to filter incorrectly collected noise images. Experiment results on Multi30K and other two translation datasets show that our proposed method achieves significant improvements over strong baselines.
2018
pdf
Automatic Identification of Indicators of Compromise using Neural-Based Sequence Labelling
Shengping Zhou
|
Zi Long
|
Lianzhi Tan
|
Hao Guo
Proceedings of the 32nd Pacific Asia Conference on Language, Information and Computation
2017
pdf
abs
Patent NMT integrated with Large Vocabulary Phrase Translation by SMT at WAT 2017
Zi Long
|
Ryuichiro Kimura
|
Takehito Utsuro
|
Tomoharu Mitsuhashi
|
Mikio Yamamoto
Proceedings of the 4th Workshop on Asian Translation (WAT2017)
Neural machine translation (NMT) cannot handle a larger vocabulary because the training complexity and decoding complexity proportionally increase with the number of target words. This problem becomes even more serious when translating patent documents, which contain many technical terms that are observed infrequently. Long et al.(2017) proposed to select phrases that contain out-of-vocabulary words using the statistical approach of branching entropy. The selected phrases are then replaced with tokens during training and post-translated by the phrase translation table of SMT. In this paper, we apply the method proposed by Long et al. (2017) to the WAT 2017 Japanese-Chinese and Japanese-English patent datasets. Evaluation on Japanese-to-Chinese, Chinese-to-Japanese, Japanese-to-English and English-to-Japanese patent sentence translation proved the effectiveness of phrases selected with branching entropy, where the NMT model of Long et al.(2017) achieves a substantial improvement over a baseline NMT model without the technique proposed by Long et al.(2017).
pdf
Neural Machine Translation Model with a Large Vocabulary Selected by Branching Entropy
Zi Long
|
Ryuichiro Kimura
|
Takehito Utsuro
|
Tomoharu Mitsuhashi
|
Mikio Yamamoto
Proceedings of Machine Translation Summit XVI: Research Track
2016
pdf
bib
abs
Translation of Patent Sentences with a Large Vocabulary of Technical Terms Using Neural Machine Translation
Zi Long
|
Takehito Utsuro
|
Tomoharu Mitsuhashi
|
Mikio Yamamoto
Proceedings of the 3rd Workshop on Asian Translation (WAT2016)
Neural machine translation (NMT), a new approach to machine translation, has achieved promising results comparable to those of traditional approaches such as statistical machine translation (SMT). Despite its recent success, NMT cannot handle a larger vocabulary because training complexity and decoding complexity proportionally increase with the number of target words. This problem becomes even more serious when translating patent documents, which contain many technical terms that are observed infrequently. In NMTs, words that are out of vocabulary are represented by a single unknown token. In this paper, we propose a method that enables NMT to translate patent sentences comprising a large vocabulary of technical terms. We train an NMT system on bilingual data wherein technical terms are replaced with technical term tokens; this allows it to translate most of the source sentences except technical terms. Further, we use it as a decoder to translate source sentences with technical term tokens and replace the tokens with technical term translations using SMT. We also use it to rerank the 1,000-best SMT translations on the basis of the average of the SMT score and that of the NMT rescoring of the translated sentences with technical term tokens. Our experiments on Japanese-Chinese patent sentences show that the proposed NMT system achieves a substantial improvement of up to 3.1 BLEU points and 2.3 RIBES points over traditional SMT systems and an improvement of approximately 0.6 BLEU points and 0.8 RIBES points over an equivalent NMT system without our proposed technique.
2015
pdf
Collecting bilingual technical terms from patent families of character-segmented Chinese sentences and morpheme-segmented Japanese sentences
Zi Long
|
Takehito Utsuro
|
Tomoharu Mitsuhashi
|
Mikio Yamamoto
Proceedings of the 6th Workshop on Patent and Scientific Literature Translation
pdf
Evaluating Features for Identifying Japanese-Chinese Bilingual Synonymous Technical Terms from Patent Families
Zi Long
|
Takehito Utsuro
|
Tomoharu Mitsuhashi
|
Mikio Yamamoto
Proceedings of the Eighth Workshop on Building and Using Comparable Corpora
2013
pdf
bib
Compositional translation of technical terms by integrating patent families as a parallel corpus and a comparable corpus
Itsuki Toyota
|
Zi Long
|
Lijuan Dong
|
Takehito Utsuro
|
Mikio Yamamoto
Proceedings of the 5th Workshop on Patent Translation