@inproceedings{labeau-allauzen-2018-learning,
title = "Learning with Noise-Contrastive Estimation: Easing training by learning to scale",
author = "Labeau, Matthieu and
Allauzen, Alexandre",
editor = "Bender, Emily M. and
Derczynski, Leon and
Isabelle, Pierre",
booktitle = "Proceedings of the 27th International Conference on Computational Linguistics",
month = aug,
year = "2018",
address = "Santa Fe, New Mexico, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/C18-1261/",
pages = "3090--3101",
abstract = "Noise-Contrastive Estimation (NCE) is a learning criterion that is regularly used to train neural language models in place of Maximum Likelihood Estimation, since it avoids the computational bottleneck caused by the output softmax. In this paper, we analyse and explain some of the weaknesses of this objective function, linked to the mechanism of self-normalization, by closely monitoring comparative experiments. We then explore several remedies and modifications to propose tractable and efficient NCE training strategies. In particular, we propose to make the scaling factor a trainable parameter of the model, and to use the noise distribution to initialize the output bias. These solutions, yet simple, yield stable and competitive performances in either small and large scale language modelling tasks."
}
Markdown (Informal)
[Learning with Noise-Contrastive Estimation: Easing training by learning to scale](https://preview.aclanthology.org/jlcl-multiple-ingestion/C18-1261/) (Labeau & Allauzen, COLING 2018)
ACL