@inproceedings{lee-bloem-2023-comparing,
title = "Comparing domain-specific and domain-general {BERT} variants for inferred real-world knowledge through rare grammatical features in {S}erbian",
author = "Lee, Sofia and
Bloem, Jelke",
editor = "Piskorski, Jakub and
Marci{\'n}czuk, Micha{\l} and
Nakov, Preslav and
Ogrodniczuk, Maciej and
Pollak, Senja and
P{\v{r}}ib{\'a}{\v{n}}, Pavel and
Rybak, Piotr and
Steinberger, Josef and
Yangarber, Roman",
booktitle = "Proceedings of the 9th Workshop on Slavic Natural Language Processing 2023 (SlavicNLP 2023)",
month = may,
year = "2023",
address = "Dubrovnik, Croatia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.bsnlp-1.7/",
doi = "10.18653/v1/2023.bsnlp-1.7",
pages = "47--60",
abstract = "Transfer learning is one of the prevailing approaches towards training language-specific BERT models. However, some languages have uncommon features that may prove to be challenging to more domain-general models but not domain-specific models. Comparing the performance of BERTi{\'c}, a Bosnian-Croatian-Montenegrin-Serbian model, and Multilingual BERT on a Named-Entity Recognition (NER) task and Masked Language Modelling (MLM) task based around a rare phenomenon of indeclinable female foreign names in Serbian reveals how the different training approaches impacts their performance. Multilingual BERT is shown to perform better than BERTi{\'c} in the NER task, but BERTi{\'c} greatly exceeds in the MLM task. Thus, there are applications both for domain-general training and domain-specific training depending on the tasks at hand."
}
Markdown (Informal)
[Comparing domain-specific and domain-general BERT variants for inferred real-world knowledge through rare grammatical features in Serbian](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.bsnlp-1.7/) (Lee & Bloem, BSNLP 2023)
ACL