@inproceedings{huang-etal-2018-large,
title = "Large Margin Neural Language Model",
author = "Huang, Jiaji and
Li, Yi and
Ping, Wei and
Huang, Liang",
editor = "Riloff, Ellen and
Chiang, David and
Hockenmaier, Julia and
Tsujii, Jun{'}ichi",
booktitle = "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing",
month = oct # "-" # nov,
year = "2018",
address = "Brussels, Belgium",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/D18-1150/",
doi = "10.18653/v1/D18-1150",
pages = "1183--1191",
abstract = "We propose a large margin criterion for training neural language models. Conventionally, neural language models are trained by minimizing perplexity (PPL) on grammatical sentences. However, we demonstrate that PPL may not be the best metric to optimize in some tasks, and further propose a large margin formulation. The proposed method aims to enlarge the margin between the ``good'' and ``bad'' sentences in a task-specific sense. It is trained end-to-end and can be widely applied to tasks that involve re-scoring of generated text. Compared with minimum-PPL training, our method gains up to 1.1 WER reduction for speech recognition and 1.0 BLEU increase for machine translation."
}
Markdown (Informal)
[Large Margin Neural Language Model](https://preview.aclanthology.org/fix-sig-urls/D18-1150/) (Huang et al., EMNLP 2018)
ACL
- Jiaji Huang, Yi Li, Wei Ping, and Liang Huang. 2018. Large Margin Neural Language Model. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 1183–1191, Brussels, Belgium. Association for Computational Linguistics.