Improving Latvian Morphosyntactic Parsing with Pretrained Encoders and Analyzer-Constrained Decoding

Arturs Znotins


Abstract
We present a systematic evaluation of Latvian morphosyntactic parsing with pretrained transformer encoders in a unified joint architecture for tagging, lemmatization, and dependency parsing. We benchmark multilingual and Latvian-specific models and show that language-specific adaptation, even with modest in-language data, substantially improves performance. We further demonstrate that factored morphological modeling improves robustness and that integrating a Latvian morphological analyzer through constrained decoding yields consistent gains in XPOS tagging and lemmatization. The best system achieves new state-of-the-art results, reaching 95.22% XPOS accuracy, 98.72% lemma accuracy, and 93.19% LAS.
Anthology ID:
2026.lrec-main.918
Volume:
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Month:
May
Year:
2026
Address:
Palma de Mallorca, Spain
Editors:
Stelios Piperidis, Núria Bel, Henk van den Heuvel, Nancy Ide, Simon Krek, Antonio Toral
Venue:
LREC
SIG:
Publisher:
ELRA Language Resource Association
Note:
Pages:
11724–11734
Language:
URL:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.918/
DOI:
Bibkey:
Cite (ACL):
Arturs Znotins. 2026. Improving Latvian Morphosyntactic Parsing with Pretrained Encoders and Analyzer-Constrained Decoding. International Conference on Language Resources and Evaluation, main:11724–11734.
Cite (Informal):
Improving Latvian Morphosyntactic Parsing with Pretrained Encoders and Analyzer-Constrained Decoding (Znotins, LREC 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-lrec/2026.lrec-main.918.pdf