@inproceedings{gupta-zhang-2018-attend,
title = "To Attend or not to Attend: A Case Study on Syntactic Structures for Semantic Relatedness",
author = "Gupta, Amulya and
Zhang, Zhu",
editor = "Gurevych, Iryna and
Miyao, Yusuke",
booktitle = "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2018",
address = "Melbourne, Australia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1197/",
doi = "10.18653/v1/P18-1197",
pages = "2116--2125",
abstract = "With the recent success of Recurrent Neural Networks (RNNs) in Machine Translation (MT), attention mechanisms have become increasingly popular. The purpose of this paper is two-fold; firstly, we propose a novel attention model on Tree Long Short-Term Memory Networks (Tree-LSTMs), a tree-structured generalization of standard LSTM. Secondly, we study the interaction between attention and syntactic structures, by experimenting with three LSTM variants: bidirectional-LSTMs, Constituency Tree-LSTMs, and Dependency Tree-LSTMs. Our models are evaluated on two semantic relatedness tasks: semantic relatedness scoring for sentence pairs (SemEval 2012, Task 6 and SemEval 2014, Task 1) and paraphrase detection for question pairs (Quora, 2017)."
}
Markdown (Informal)
[To Attend or not to Attend: A Case Study on Syntactic Structures for Semantic Relatedness](https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1197/) (Gupta & Zhang, ACL 2018)
ACL