@inproceedings{bhat-etal-2019-margin,
title = "A Margin-based Loss with Synthetic Negative Samples for Continuous-output Machine Translation",
author = "Bhat, Gayatri and
Kumar, Sachin and
Tsvetkov, Yulia",
editor = "Birch, Alexandra and
Finch, Andrew and
Hayashi, Hiroaki and
Konstas, Ioannis and
Luong, Thang and
Neubig, Graham and
Oda, Yusuke and
Sudoh, Katsuhito",
booktitle = "Proceedings of the 3rd Workshop on Neural Generation and Translation",
month = nov,
year = "2019",
address = "Hong Kong",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/D19-5621/",
doi = "10.18653/v1/D19-5621",
pages = "199--205",
abstract = "Neural models that eliminate the softmax bottleneck by generating word embeddings (rather than multinomial distributions over a vocabulary) attain faster training with fewer learnable parameters. These models are currently trained by maximizing densities of pretrained target embeddings under von Mises-Fisher distributions parameterized by corresponding model-predicted embeddings. This work explores the utility of margin-based loss functions in optimizing such models. We present syn-margin loss, a novel margin-based loss that uses a synthetic negative sample constructed from only the predicted and target embeddings at every step. The loss is efficient to compute, and we use a geometric analysis to argue that it is more consistent and interpretable than other margin-based losses. Empirically, we find that syn-margin provides small but significant improvements over both vMF and standard margin-based losses in continuous-output neural machine translation."
}
Markdown (Informal)
[A Margin-based Loss with Synthetic Negative Samples for Continuous-output Machine Translation](https://preview.aclanthology.org/jlcl-multiple-ingestion/D19-5621/) (Bhat et al., NGT 2019)
ACL