@inproceedings{salton-etal-2017-attentive,
title = "Attentive Language Models",
author = "Salton, Giancarlo and
Ross, Robert and
Kelleher, John",
editor = "Kondrak, Greg and
Watanabe, Taro",
booktitle = "Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers)",
month = nov,
year = "2017",
address = "Taipei, Taiwan",
publisher = "Asian Federation of Natural Language Processing",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/I17-1045/",
pages = "441--450",
abstract = "In this paper, we extend Recurrent Neural Network Language Models (RNN-LMs) with an attention mechanism. We show that an {\textquotedblleft}attentive{\textquotedblright} RNN-LM (with 11M parameters) achieves a better perplexity than larger RNN-LMs (with 66M parameters) and achieves performance comparable to an ensemble of 10 similar sized RNN-LMs. We also show that an {\textquotedblleft}attentive{\textquotedblright} RNN-LM needs less contextual information to achieve similar results to the state-of-the-art on the wikitext2 dataset."
}
Markdown (Informal)
[Attentive Language Models](https://preview.aclanthology.org/add-emnlp-2024-awards/I17-1045/) (Salton et al., IJCNLP 2017)
ACL
- Giancarlo Salton, Robert Ross, and John Kelleher. 2017. Attentive Language Models. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 441–450, Taipei, Taiwan. Asian Federation of Natural Language Processing.