@inproceedings{yang-etal-2019-using,
title = "Using Large Corpus N-gram Statistics to Improve Recurrent Neural Language Models",
author = "Yang, Yiben and
Wang, Ji-Ping and
Downey, Doug",
editor = "Burstein, Jill and
Doran, Christy and
Solorio, Thamar",
booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
month = jun,
year = "2019",
address = "Minneapolis, Minnesota",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/N19-1330/",
doi = "10.18653/v1/N19-1330",
pages = "3268--3273",
abstract = "Recurrent neural network language models (RNNLM) form a valuable foundation for many NLP systems, but training the models can be computationally expensive, and may take days to train on a large corpus. We explore a technique that uses large corpus n-gram statistics as a regularizer for training a neural network LM on a smaller corpus. In experiments with the Billion-Word and Wikitext corpora, we show that the technique is effective, and more time-efficient than simply training on a larger sequential corpus. We also introduce new strategies for selecting the most informative n-grams, and show that these boost efficiency."
}
Markdown (Informal)
[Using Large Corpus N-gram Statistics to Improve Recurrent Neural Language Models](https://preview.aclanthology.org/fix-sig-urls/N19-1330/) (Yang et al., NAACL 2019)
ACL