UZH@CRAFT-ST: a Sequence-labeling Approach to Concept Recognition

Lenz Furrer, Joseph Cornelius, Fabio Rinaldi


Abstract
As our submission to the CRAFT shared task 2019, we present two neural approaches to concept recognition. We propose two different systems for joint named entity recognition (NER) and normalization (NEN), both of which model the task as a sequence labeling problem. Our first system is a BiLSTM network with two separate outputs for NER and NEN trained from scratch, whereas the second system is an instance of BioBERT fine-tuned on the concept-recognition task. We exploit two strategies for extending concept coverage, ontology pretraining and backoff with a dictionary lookup. Our results show that the backoff strategy effectively tackles the problem of unseen concepts, addressing a major limitation of the chosen design. In the cross-system comparison, BioBERT proves to be a strong basis for creating a concept-recognition system, although some entity types are predicted more accurately by the BiLSTM-based system.
Anthology ID:
D19-5726
Volume:
Proceedings of the 5th Workshop on BioNLP Open Shared Tasks
Month:
November
Year:
2019
Address:
Hong Kong, China
Venue:
BioNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
185–195
Language:
URL:
https://aclanthology.org/D19-5726
DOI:
10.18653/v1/D19-5726
Bibkey:
Cite (ACL):
Lenz Furrer, Joseph Cornelius, and Fabio Rinaldi. 2019. UZH@CRAFT-ST: a Sequence-labeling Approach to Concept Recognition. In Proceedings of the 5th Workshop on BioNLP Open Shared Tasks, pages 185–195, Hong Kong, China. Association for Computational Linguistics.
Cite (Informal):
UZH@CRAFT-ST: a Sequence-labeling Approach to Concept Recognition (Furrer et al., BioNLP 2019)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-script-update/D19-5726.pdf