Named Entity Recognition with Partially Annotated Training Data

Stephen Mayhew, Snigdha Chaturvedi, Chen-Tse Tsai, Dan Roth


Abstract
Supervised machine learning assumes the availability of fully-labeled data, but in many cases, such as low-resource languages, the only data available is partially annotated. We study the problem of Named Entity Recognition (NER) with partially annotated training data in which a fraction of the named entities are labeled, and all other tokens, entities or otherwise, are labeled as non-entity by default. In order to train on this noisy dataset, we need to distinguish between the true and false negatives. To this end, we introduce a constraint-driven iterative algorithm that learns to detect false negatives in the noisy set and downweigh them, resulting in a weighted training set. With this set, we train a weighted NER model. We evaluate our algorithm with weighted variants of neural and non-neural NER models on data in 8 languages from several language and script families, showing strong ability to learn from partial data. Finally, to show real-world efficacy, we evaluate on a Bengali NER corpus annotated by non-speakers, outperforming the prior state-of-the-art by over 5 points F1.
Anthology ID:
K19-1060
Volume:
Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)
Month:
November
Year:
2019
Address:
Hong Kong, China
Editors:
Mohit Bansal, Aline Villavicencio
Venue:
CoNLL
SIG:
SIGNLL
Publisher:
Association for Computational Linguistics
Note:
Pages:
645–655
Language:
URL:
https://aclanthology.org/K19-1060
DOI:
10.18653/v1/K19-1060
Bibkey:
Cite (ACL):
Stephen Mayhew, Snigdha Chaturvedi, Chen-Tse Tsai, and Dan Roth. 2019. Named Entity Recognition with Partially Annotated Training Data. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 645–655, Hong Kong, China. Association for Computational Linguistics.
Cite (Informal):
Named Entity Recognition with Partially Annotated Training Data (Mayhew et al., CoNLL 2019)
Copy Citation:
PDF:
https://preview.aclanthology.org/emnlp-22-attachments/K19-1060.pdf