@inproceedings{kassner-etal-2021-multilingual,
title = "Multilingual {LAMA}: Investigating Knowledge in Multilingual Pretrained Language Models",
author = {Kassner, Nora and
Dufter, Philipp and
Sch{\"u}tze, Hinrich},
editor = "Merlo, Paola and
Tiedemann, Jorg and
Tsarfaty, Reut",
booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume",
month = apr,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2021.eacl-main.284/",
doi = "10.18653/v1/2021.eacl-main.284",
pages = "3250--3258",
abstract = "Recently, it has been found that monolingual English language models can be used as knowledge bases. Instead of structural knowledge base queries, masked sentences such as ``Paris is the capital of [MASK]'' are used as probes. We translate the established benchmarks TREx and GoogleRE into 53 languages. Working with mBERT, we investigate three questions. (i) Can mBERT be used as a multilingual knowledge base? Most prior work only considers English. Extending research to multiple languages is important for diversity and accessibility. (ii) Is mBERT{'}s performance as knowledge base language-independent or does it vary from language to language? (iii) A multilingual model is trained on more text, e.g., mBERT is trained on 104 Wikipedias. Can mBERT leverage this for better performance? We find that using mBERT as a knowledge base yields varying performance across languages and pooling predictions across languages improves performance. Conversely, mBERT exhibits a language bias; e.g., when queried in Italian, it tends to predict Italy as the country of origin."
}
Markdown (Informal)
[Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models](https://preview.aclanthology.org/fix-sig-urls/2021.eacl-main.284/) (Kassner et al., EACL 2021)
ACL