Auxiliary Learning for Named Entity Recognition with Multiple Auxiliary Biomedical Training Data

Taiki Watanabe, Tomoya Ichikawa, Akihiro Tamura, Tomoya Iwakura, Chunpeng Ma, Tsuneo Kato


Abstract
Named entity recognition (NER) is one of the elemental technologies, which has been used for knowledge extraction from biomedical text. As one of the NER improvement approaches, multi-task learning that learns a model from multiple training data has been used. Among multi-task learning, an auxiliary learning method, which uses an auxiliary task for improving its target task, has shown higher NER performance than conventional multi-task learning for improving all the tasks simultaneously by using only one auxiliary task in the auxiliary learning. We propose Multiple Utilization of NER Corpora Helpful for Auxiliary BLESsing (MUNCH ABLES). MUNCHABLES utilizes multiple training datasets as auxiliary training data by the following methods; the first one is to finetune the NER model of the target task by sequentially performing auxiliary learning for each auxiliary training dataset, and the other is to use all training datasets in one auxiliary learning. We evaluate MUNCHABLES on eight biomedical-related domain NER tasks, where seven training datasets are used as auxiliary training data. The experiment results show that MUNCHABLES achieves higher accuracy than conventional multi-task learning methods on average while showing state-of-the-art accuracy.
Anthology ID:
2022.bionlp-1.13
Volume:
Proceedings of the 21st Workshop on Biomedical Language Processing
Month:
May
Year:
2022
Address:
Dublin, Ireland
Venue:
BioNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
130–139
Language:
URL:
https://aclanthology.org/2022.bionlp-1.13
DOI:
10.18653/v1/2022.bionlp-1.13
Bibkey:
Cite (ACL):
Taiki Watanabe, Tomoya Ichikawa, Akihiro Tamura, Tomoya Iwakura, Chunpeng Ma, and Tsuneo Kato. 2022. Auxiliary Learning for Named Entity Recognition with Multiple Auxiliary Biomedical Training Data. In Proceedings of the 21st Workshop on Biomedical Language Processing, pages 130–139, Dublin, Ireland. Association for Computational Linguistics.
Cite (Informal):
Auxiliary Learning for Named Entity Recognition with Multiple Auxiliary Biomedical Training Data (Watanabe et al., BioNLP 2022)
Copy Citation:
PDF:
https://preview.aclanthology.org/paclic-22-ingestion/2022.bionlp-1.13.pdf
Video:
 https://preview.aclanthology.org/paclic-22-ingestion/2022.bionlp-1.13.mp4
Data
NCBI Disease