@inproceedings{gupta-etal-2017-entity,
    title = "Entity Linking via Joint Encoding of Types, Descriptions, and Context",
    author = "Gupta, Nitish  and
      Singh, Sameer  and
      Roth, Dan",
    editor = "Palmer, Martha  and
      Hwa, Rebecca  and
      Riedel, Sebastian",
    booktitle = "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing",
    month = sep,
    year = "2017",
    address = "Copenhagen, Denmark",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/iwcs-25-ingestion/D17-1284/",
    doi = "10.18653/v1/D17-1284",
    pages = "2681--2690",
    abstract = "For accurate entity linking, we need to capture various information aspects of an entity, such as its description in a KB, contexts in which it is mentioned, and structured knowledge. Additionally, a linking system should work on texts from different domains without requiring domain-specific training data or hand-engineered features. In this work we present a neural, modular entity linking system that learns a unified dense representation for each entity using multiple sources of information, such as its description, contexts around its mentions, and its fine-grained types. We show that the resulting entity linking system is effective at combining these sources, and performs competitively, sometimes out-performing current state-of-the-art systems across datasets, without requiring any domain-specific training data or hand-engineered features. We also show that our model can effectively ``embed'' entities that are new to the KB, and is able to link its mentions accurately."
}Markdown (Informal)
[Entity Linking via Joint Encoding of Types, Descriptions, and Context](https://preview.aclanthology.org/iwcs-25-ingestion/D17-1284/) (Gupta et al., EMNLP 2017)
ACL