@inproceedings{melamud-etal-2017-simple,
title = "A Simple Language Model based on {PMI} Matrix Approximations",
author = "Melamud, Oren and
Dagan, Ido and
Goldberger, Jacob",
editor = "Palmer, Martha and
Hwa, Rebecca and
Riedel, Sebastian",
booktitle = "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing",
month = sep,
year = "2017",
address = "Copenhagen, Denmark",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/D17-1198/",
doi = "10.18653/v1/D17-1198",
pages = "1860--1865",
abstract = "In this study, we introduce a new approach for learning language models by training them to estimate word-context pointwise mutual information (PMI), and then deriving the desired conditional probabilities from PMI at test time. Specifically, we show that with minor modifications to word2vec{'}s algorithm, we get principled language models that are closely related to the well-established Noise Contrastive Estimation (NCE) based language models. A compelling aspect of our approach is that our models are trained with the same simple negative sampling objective function that is commonly used in word2vec to learn word embeddings."
}
Markdown (Informal)
[A Simple Language Model based on PMI Matrix Approximations](https://preview.aclanthology.org/fix-sig-urls/D17-1198/) (Melamud et al., EMNLP 2017)
ACL