Abstract
In this work we propose a novel end-to-end multi-stage Knowledge Graph (KG) generation system from textual inputs, separating the overall process into two stages. The graph nodes are generated first using pretrained language model, followed by a simple edge construction head, enabling efficient KG extraction from the text. For each stage we consider several architectural choices that can be used depending on the available training resources. We evaluated the model on a recent WebNLG 2020 Challenge dataset, matching the state-of-the-art performance on text-to-RDF generation task, as well as on New York Times (NYT) and a large-scale TekGen datasets, showing strong overall performance, outperforming the existing baselines. We believe that the proposed system can serve as a viable KG construction alternative to the existing linearization or sampling-based graph generation approaches.- Anthology ID:
- 2022.findings-emnlp.116
- Volume:
- Findings of the Association for Computational Linguistics: EMNLP 2022
- Month:
- December
- Year:
- 2022
- Address:
- Abu Dhabi, United Arab Emirates
- Editors:
- Yoav Goldberg, Zornitsa Kozareva, Yue Zhang
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 1610–1622
- Language:
- URL:
- https://aclanthology.org/2022.findings-emnlp.116
- DOI:
- 10.18653/v1/2022.findings-emnlp.116
- Cite (ACL):
- Igor Melnyk, Pierre Dognin, and Payel Das. 2022. Knowledge Graph Generation From Text. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 1610–1622, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
- Cite (Informal):
- Knowledge Graph Generation From Text (Melnyk et al., Findings 2022)
- PDF:
- https://preview.aclanthology.org/naacl24-info/2022.findings-emnlp.116.pdf