@inproceedings{bhat-etal-2019-margin,
    title = "A Margin-based Loss with Synthetic Negative Samples for Continuous-output Machine Translation",
    author = "Bhat, Gayatri  and
      Kumar, Sachin  and
      Tsvetkov, Yulia",
    editor = "Birch, Alexandra  and
      Finch, Andrew  and
      Hayashi, Hiroaki  and
      Konstas, Ioannis  and
      Luong, Thang  and
      Neubig, Graham  and
      Oda, Yusuke  and
      Sudoh, Katsuhito",
    booktitle = "Proceedings of the 3rd Workshop on Neural Generation and Translation",
    month = nov,
    year = "2019",
    address = "Hong Kong",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/iwcs-25-ingestion/D19-5621/",
    doi = "10.18653/v1/D19-5621",
    pages = "199--205",
    abstract = "Neural models that eliminate the softmax bottleneck by generating word embeddings (rather than multinomial distributions over a vocabulary) attain faster training with fewer learnable parameters. These models are currently trained by maximizing densities of pretrained target embeddings under von Mises-Fisher distributions parameterized by corresponding model-predicted embeddings. This work explores the utility of margin-based loss functions in optimizing such models. We present syn-margin loss, a novel margin-based loss that uses a synthetic negative sample constructed from only the predicted and target embeddings at every step. The loss is efficient to compute, and we use a geometric analysis to argue that it is more consistent and interpretable than other margin-based losses. Empirically, we find that syn-margin provides small but significant improvements over both vMF and standard margin-based losses in continuous-output neural machine translation."
}Markdown (Informal)
[A Margin-based Loss with Synthetic Negative Samples for Continuous-output Machine Translation](https://preview.aclanthology.org/iwcs-25-ingestion/D19-5621/) (Bhat et al., NGT 2019)
ACL