Abstract
Conceptual spaces are geometric representations of meaning that were proposed by G ̈ardenfors (2000). They share many similarities with the vector space embeddings that are commonly used in natural language processing. However, rather than representing entities in a single vector space, conceptual spaces are usually decomposed into several facets, each of which is then modelled as a relatively low dimensional vector space. Unfortunately, the problem of learning such conceptual spaces has thus far only received limited attention. To address this gap, we analyze how, and to what extent, a given vector space embedding can be decomposed into meaningful facets in an unsupervised fashion. While this problem is highly challenging, we show that useful facets can be discovered by relying on word embeddings to group semantically related features.- Anthology ID:
- K19-1013
- Volume:
- Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)
- Month:
- November
- Year:
- 2019
- Address:
- Hong Kong, China
- Editors:
- Mohit Bansal, Aline Villavicencio
- Venue:
- CoNLL
- SIG:
- SIGNLL
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 131–139
- Language:
- URL:
- https://aclanthology.org/K19-1013
- DOI:
- 10.18653/v1/K19-1013
- Cite (ACL):
- Rana Alshaikh, Zied Bouraoui, and Steven Schockaert. 2019. Learning Conceptual Spaces with Disentangled Facets. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 131–139, Hong Kong, China. Association for Computational Linguistics.
- Cite (Informal):
- Learning Conceptual Spaces with Disentangled Facets (Alshaikh et al., CoNLL 2019)
- PDF:
- https://preview.aclanthology.org/improve-issue-templates/K19-1013.pdf