Abstract
This paper introduces our method in the system for SemEval 2023 Task 2: MultiCoNER II Multilingual Complex Named Entity Recognition, Track 9-Chinese. This task focuses on detecting fine-grained named entities whose data set has a fine-grained taxonomy of 36 NE classes, representing a realistic challenge for NER. In this task, we need to identify entity boundaries and category labels for the six identified categories. We use BERT embedding to represent each character in the original sentence and train CRF-Rdrop to predict named entity categories using the data set provided by the organizer. Our best submission, with a macro average F1 score of 0.5657, ranked 15th out of 22 teams.- Anthology ID:
- 2023.semeval-1.224
- Volume:
- Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023)
- Month:
- July
- Year:
- 2023
- Address:
- Toronto, Canada
- Venue:
- SemEval
- SIG:
- SIGLEX
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 1619–1624
- Language:
- URL:
- https://aclanthology.org/2023.semeval-1.224
- DOI:
- 10.18653/v1/2023.semeval-1.224
- Cite (ACL):
- Jing Li and Xiaobing Zhou. 2023. YNUNLP at SemEval-2023 Task 2: The Pseudo Twin Tower Pre-training Model for Chinese Named Entity Recognition. In Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), pages 1619–1624, Toronto, Canada. Association for Computational Linguistics.
- Cite (Informal):
- YNUNLP at SemEval-2023 Task 2: The Pseudo Twin Tower Pre-training Model for Chinese Named Entity Recognition (Li & Zhou, SemEval 2023)
- PDF:
- https://preview.aclanthology.org/remove-xml-comments/2023.semeval-1.224.pdf