Abstract
Having access to high-quality grammatical annotations is important for downstream tasks in NLP as well as for corpus-based research. In this paper, we describe experiments with the Latin BERT word embeddings that were recently be made available by Bamman and Burns (2020). We show that these embeddings produce competitive results in the low-level task morpho-syntactic tagging. In addition, we describe a graph-based dependency parser that is trained with these embeddings and that clearly outperforms various baselines.- Anthology ID:
- 2022.lt4hala-1.3
- Volume:
- Proceedings of the Second Workshop on Language Technologies for Historical and Ancient Languages
- Month:
- June
- Year:
- 2022
- Address:
- Marseille, France
- Editors:
- Rachele Sprugnoli, Marco Passarotti
- Venue:
- LT4HALA
- SIG:
- Publisher:
- European Language Resources Association
- Note:
- Pages:
- 20–25
- Language:
- URL:
- https://aclanthology.org/2022.lt4hala-1.3
- DOI:
- Cite (ACL):
- Sebastian Nehrdich and Oliver Hellwig. 2022. Accurate Dependency Parsing and Tagging of Latin. In Proceedings of the Second Workshop on Language Technologies for Historical and Ancient Languages, pages 20–25, Marseille, France. European Language Resources Association.
- Cite (Informal):
- Accurate Dependency Parsing and Tagging of Latin (Nehrdich & Hellwig, LT4HALA 2022)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-4/2022.lt4hala-1.3.pdf