Abstract
Implicit discourse relation recognition is an extremely challenging task due to the lack of indicative connectives. Various neural network architectures have been proposed for this task recently, but most of them suffer from the shortage of labeled data. In this paper, we address this problem by procuring additional training data from parallel corpora: When humans translate a text, they sometimes add connectives (a process known as explicitation). We automatically back-translate it into an English connective and use it to infer a label with high confidence. We show that a training set several times larger than the original training set can be generated this way. With the extra labeled instances, we show that even a simple bidirectional Long Short-Term Memory Network can outperform the current state-of-the-art.- Anthology ID:
- I17-1049
- Volume:
- Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
- Month:
- November
- Year:
- 2017
- Address:
- Taipei, Taiwan
- Editors:
- Greg Kondrak, Taro Watanabe
- Venue:
- IJCNLP
- SIG:
- Publisher:
- Asian Federation of Natural Language Processing
- Note:
- Pages:
- 484–495
- Language:
- URL:
- https://aclanthology.org/I17-1049
- DOI:
- Cite (ACL):
- Wei Shi, Frances Yung, Raphael Rubino, and Vera Demberg. 2017. Using Explicit Discourse Connectives in Translation for Implicit Discourse Relation Classification. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 484–495, Taipei, Taiwan. Asian Federation of Natural Language Processing.
- Cite (Informal):
- Using Explicit Discourse Connectives in Translation for Implicit Discourse Relation Classification (Shi et al., IJCNLP 2017)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-5/I17-1049.pdf