@inproceedings{kim-etal-2017-unsupervised,
title = "Unsupervised Training for Large Vocabulary Translation Using Sparse Lexicon and Word Classes",
author = "Kim, Yunsu and
Schamper, Julian and
Ney, Hermann",
editor = "Lapata, Mirella and
Blunsom, Phil and
Koller, Alexander",
booktitle = "Proceedings of the 15th Conference of the {E}uropean Chapter of the Association for Computational Linguistics: Volume 2, Short Papers",
month = apr,
year = "2017",
address = "Valencia, Spain",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/E17-2103/",
pages = "650--656",
abstract = "We address for the first time unsupervised training for a translation task with hundreds of thousands of vocabulary words. We scale up the expectation-maximization (EM) algorithm to learn a large translation table without any parallel text or seed lexicon. First, we solve the memory bottleneck and enforce the sparsity with a simple thresholding scheme for the lexicon. Second, we initialize the lexicon training with word classes, which efficiently boosts the performance. Our methods produced promising results on two large-scale unsupervised translation tasks."
}
Markdown (Informal)
[Unsupervised Training for Large Vocabulary Translation Using Sparse Lexicon and Word Classes](https://preview.aclanthology.org/fix-sig-urls/E17-2103/) (Kim et al., EACL 2017)
ACL