Deep Span Representations for Named Entity Recognition

Enwei Zhu, Yiyang Liu, Jinpeng Li


Abstract
Span-based models are one of the most straightforward methods for named entity recognition (NER). Existing span-based NER systems shallowly aggregate the token representations to span representations. However, this typically results in significant ineffectiveness for long entities, a coupling between the representations of overlapping spans, and ultimately a performance degradation. In this study, we propose DSpERT (Deep Span Encoder Representations from Transformers), which comprises a standard Transformer and a span Transformer. The latter uses low-layered span representations as queries, and aggregates the token representations as keys and values, layer by layer from bottom to top. Thus, DSpERT produces span representations of deep semantics. With weight initialization from pretrained language models, DSpERT achieves performance higher than or competitive with recent state-of-the-art systems on six NER benchmarks. Experimental results verify the importance of the depth for span representations, and show that DSpERT performs particularly well on long-span entities and nested structures. Further, the deep span representations are well structured and easily separable in the feature space.
Anthology ID:
2023.findings-acl.672
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
10565–10582
Language:
URL:
https://aclanthology.org/2023.findings-acl.672
DOI:
10.18653/v1/2023.findings-acl.672
Bibkey:
Cite (ACL):
Enwei Zhu, Yiyang Liu, and Jinpeng Li. 2023. Deep Span Representations for Named Entity Recognition. In Findings of the Association for Computational Linguistics: ACL 2023, pages 10565–10582, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Deep Span Representations for Named Entity Recognition (Zhu et al., Findings 2023)
Copy Citation:
PDF:
https://preview.aclanthology.org/dois-2013-emnlp/2023.findings-acl.672.pdf