@inproceedings{nandakumar-etal-2019-well,
    title = "How Well Do Embedding Models Capture Non-compositionality? A View from Multiword Expressions",
    author = "Nandakumar, Navnita  and
      Baldwin, Timothy  and
      Salehi, Bahar",
    editor = "Rogers, Anna  and
      Drozd, Aleksandr  and
      Rumshisky, Anna  and
      Goldberg, Yoav",
    booktitle = "Proceedings of the 3rd Workshop on Evaluating Vector Space Representations for {NLP}",
    month = jun,
    year = "2019",
    address = "Minneapolis, USA",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/W19-2004/",
    doi = "10.18653/v1/W19-2004",
    pages = "27--34",
    abstract = "In this paper, we apply various embedding methods on multiword expressions to study how well they capture the nuances of non-compositional data. Our results from a pool of word-, character-, and document-level embbedings suggest that Word2vec performs the best, followed by FastText and Infersent. Moreover, we find that recently-proposed contextualised embedding models such as Bert and ELMo are not adept at handling non-compositionality in multiword expressions."
}Markdown (Informal)
[How Well Do Embedding Models Capture Non-compositionality? A View from Multiword Expressions](https://preview.aclanthology.org/ingest-emnlp/W19-2004/) (Nandakumar et al., RepEval 2019)
ACL