Learning Nested Named Entity Recognition from Flat Annotations

Igor Rozhkov, Natalia V Loukachevitch


Abstract
Nested named entity recognition identifies entities contained within other entities, but requires expensive multi-level annotation. While flat NER corpora exist abundantly, nested resources remain scarce. We investigate whether models can learn nested structure from flat annotations alone, evaluating four approaches: string inclusions (substring matching), entity corruption (pseudo-nested data), flat neutralization (reducing false negative signal), and a hybrid fine-tuned + LLM pipeline. On NEREL, a Russian benchmark with 29 entity types where 21% of entities are nested, our best combined method achieves 26.37% inner F1, closing 40% of the gap to full nested supervision. Code is available at https://github.com/fulstock/Learning-from-Flat-Annotations.
Anthology ID:
2026.eacl-srw.50
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 4: Student Research Workshop)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Selene Baez Santamaria, Sai Ashish Somayajula, Atsuki Yamaguchi
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
649–663
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-srw.50/
DOI:
Bibkey:
Cite (ACL):
Igor Rozhkov and Natalia V Loukachevitch. 2026. Learning Nested Named Entity Recognition from Flat Annotations. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 4: Student Research Workshop), pages 649–663, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Learning Nested Named Entity Recognition from Flat Annotations (Rozhkov & Loukachevitch, EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-srw.50.pdf