Neural Tensor Networks with Diagonal Slice Matrices
Takahiro Ishihara, Katsuhiko Hayashi, Hitoshi Manabe, Masashi Shimbo, Masaaki Nagata
Abstract
Although neural tensor networks (NTNs) have been successful in many NLP tasks, they require a large number of parameters to be estimated, which often leads to overfitting and a long training time. We address these issues by applying eigendecomposition to each slice matrix of a tensor to reduce its number of paramters. First, we evaluate our proposed NTN models on knowledge graph completion. Second, we extend the models to recursive NTNs (RNTNs) and evaluate them on logical reasoning tasks. These experiments show that our proposed models learn better and faster than the original (R)NTNs.- Anthology ID:
- N18-1047
- Volume:
- Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers)
- Month:
- June
- Year:
- 2018
- Address:
- New Orleans, Louisiana
- Editors:
- Marilyn Walker, Heng Ji, Amanda Stent
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 506–515
- Language:
- URL:
- https://aclanthology.org/N18-1047
- DOI:
- 10.18653/v1/N18-1047
- Cite (ACL):
- Takahiro Ishihara, Katsuhiko Hayashi, Hitoshi Manabe, Masashi Shimbo, and Masaaki Nagata. 2018. Neural Tensor Networks with Diagonal Slice Matrices. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 506–515, New Orleans, Louisiana. Association for Computational Linguistics.
- Cite (Informal):
- Neural Tensor Networks with Diagonal Slice Matrices (Ishihara et al., NAACL 2018)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-3/N18-1047.pdf