Abstract
We present an approach to minimally supervised relation extraction that combines the benefits of learned representations and structured learning, and accurately predicts sentence-level relation mentions given only proposition-level supervision from a KB. By explicitly reasoning about missing data during learning, our approach enables large-scale training of 1D convolutional neural networks while mitigating the issue of label noise inherent in distant supervision. Our approach achieves state-of-the-art results on minimally supervised sentential relation extraction, outperforming a number of baselines, including a competitive approach that uses the attention layer of a purely neural model.- Anthology ID:
- N19-1310
- Volume:
- Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)
- Month:
- June
- Year:
- 2019
- Address:
- Minneapolis, Minnesota
- Editors:
- Jill Burstein, Christy Doran, Thamar Solorio
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 3057–3069
- Language:
- URL:
- https://aclanthology.org/N19-1310
- DOI:
- 10.18653/v1/N19-1310
- Cite (ACL):
- Fan Bai and Alan Ritter. 2019. Structured Minimally Supervised Learning for Neural Relation Extraction. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 3057–3069, Minneapolis, Minnesota. Association for Computational Linguistics.
- Cite (Informal):
- Structured Minimally Supervised Learning for Neural Relation Extraction (Bai & Ritter, NAACL 2019)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-1/N19-1310.pdf
- Code
- bflashcp3f/PCNN-NMAR