Abstract
As our submission to the CRAFT shared task 2019, we present two neural approaches to concept recognition. We propose two different systems for joint named entity recognition (NER) and normalization (NEN), both of which model the task as a sequence labeling problem. Our first system is a BiLSTM network with two separate outputs for NER and NEN trained from scratch, whereas the second system is an instance of BioBERT fine-tuned on the concept-recognition task. We exploit two strategies for extending concept coverage, ontology pretraining and backoff with a dictionary lookup. Our results show that the backoff strategy effectively tackles the problem of unseen concepts, addressing a major limitation of the chosen design. In the cross-system comparison, BioBERT proves to be a strong basis for creating a concept-recognition system, although some entity types are predicted more accurately by the BiLSTM-based system.- Anthology ID:
- D19-5726
- Volume:
- Proceedings of the 5th Workshop on BioNLP Open Shared Tasks
- Month:
- November
- Year:
- 2019
- Address:
- Hong Kong, China
- Editors:
- Kim Jin-Dong, Nédellec Claire, Bossy Robert, Deléger Louise
- Venue:
- BioNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 185–195
- Language:
- URL:
- https://aclanthology.org/D19-5726
- DOI:
- 10.18653/v1/D19-5726
- Cite (ACL):
- Lenz Furrer, Joseph Cornelius, and Fabio Rinaldi. 2019. UZH@CRAFT-ST: a Sequence-labeling Approach to Concept Recognition. In Proceedings of the 5th Workshop on BioNLP Open Shared Tasks, pages 185–195, Hong Kong, China. Association for Computational Linguistics.
- Cite (Informal):
- UZH@CRAFT-ST: a Sequence-labeling Approach to Concept Recognition (Furrer et al., BioNLP 2019)
- PDF:
- https://preview.aclanthology.org/fix-dup-bibkey/D19-5726.pdf