Abstract
Multilingual learning for Neural Named Entity Recognition (NNER) involves jointly training a neural network for multiple languages. Typically, the goal is improving the NER performance of one of the languages (the primary language) using the other assisting languages. We show that the divergence in the tag distributions of the common named entities between the primary and assisting languages can reduce the effectiveness of multilingual learning. To alleviate this problem, we propose a metric based on symmetric KL divergence to filter out the highly divergent training instances in the assisting language. We empirically show that our data selection strategy improves NER performance in many languages, including those with very limited training data.- Anthology ID:
- P18-2064
- Volume:
- Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
- Month:
- July
- Year:
- 2018
- Address:
- Melbourne, Australia
- Editors:
- Iryna Gurevych, Yusuke Miyao
- Venue:
- ACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 401–406
- Language:
- URL:
- https://aclanthology.org/P18-2064
- DOI:
- 10.18653/v1/P18-2064
- Cite (ACL):
- Rudra Murthy, Anoop Kunchukuttan, and Pushpak Bhattacharyya. 2018. Judicious Selection of Training Data in Assisting Language for Multilingual Neural NER. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 401–406, Melbourne, Australia. Association for Computational Linguistics.
- Cite (Informal):
- Judicious Selection of Training Data in Assisting Language for Multilingual Neural NER (Murthy et al., ACL 2018)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-3/P18-2064.pdf
- Code
- murthyrudra/NeuralNER