Perform like an Engine: A Closed-Loop Neural-Symbolic Learning Framework for Knowledge Graph Inference

Guanglin Niu, Bo Li, Yongfei Zhang, Shiliang Pu


Abstract
Knowledge graph (KG) inference aims to address the natural incompleteness of KGs, including rule learning-based and KG embedding (KGE) models. However, the rule learning-based models suffer from low efficiency and generalization while KGE models lack interpretability. To address these challenges, we propose a novel and effective closed-loop neural-symbolic learning framework EngineKG via incorporating our developed KGE and rule learning modules. KGE module exploits symbolic rules and paths to enhance the semantic association between entities and relations for improving KG embeddings and interpretability. A novel rule pruning mechanism is proposed in the rule learning module by leveraging paths as initial candidate rules and employing KG embeddings together with concepts for extracting more high-quality rules. Experimental results on four real-world datasets show that our model outperforms the relevant baselines on link prediction tasks, demonstrating the superiority of our KG inference model in a neural-symbolic learning fashion. The source code and datasets of this paper are available at https://github.com/ngl567/EngineKG.
Anthology ID:
2022.coling-1.119
Volume:
Proceedings of the 29th International Conference on Computational Linguistics
Month:
October
Year:
2022
Address:
Gyeongju, Republic of Korea
Editors:
Nicoletta Calzolari, Chu-Ren Huang, Hansaem Kim, James Pustejovsky, Leo Wanner, Key-Sun Choi, Pum-Mo Ryu, Hsin-Hsi Chen, Lucia Donatelli, Heng Ji, Sadao Kurohashi, Patrizia Paggio, Nianwen Xue, Seokhwan Kim, Younggyun Hahm, Zhong He, Tony Kyungil Lee, Enrico Santus, Francis Bond, Seung-Hoon Na
Venue:
COLING
SIG:
Publisher:
International Committee on Computational Linguistics
Note:
Pages:
1391–1400
Language:
URL:
https://aclanthology.org/2022.coling-1.119
DOI:
Bibkey:
Cite (ACL):
Guanglin Niu, Bo Li, Yongfei Zhang, and Shiliang Pu. 2022. Perform like an Engine: A Closed-Loop Neural-Symbolic Learning Framework for Knowledge Graph Inference. In Proceedings of the 29th International Conference on Computational Linguistics, pages 1391–1400, Gyeongju, Republic of Korea. International Committee on Computational Linguistics.
Cite (Informal):
Perform like an Engine: A Closed-Loop Neural-Symbolic Learning Framework for Knowledge Graph Inference (Niu et al., COLING 2022)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-bitext-workshop/2022.coling-1.119.pdf
Data
FB15k-237NELL-995