@inproceedings{hercig-kral-2021-evaluation,
title = "Evaluation Datasets for Cross-lingual Semantic Textual Similarity",
author = "Hercig, Tom{\'a}{\v{s}} and
Kral, Pavel",
editor = "Mitkov, Ruslan and
Angelova, Galia",
booktitle = "Proceedings of the International Conference on Recent Advances in Natural Language Processing (RANLP 2021)",
month = sep,
year = "2021",
address = "Held Online",
publisher = "INCOMA Ltd.",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.ranlp-1.59/",
pages = "524--529",
abstract = "Semantic textual similarity (STS) systems estimate the degree of the meaning similarity between two sentences. Cross-lingual STS systems estimate the degree of the meaning similarity between two sentences, each in a different language. State-of-the-art algorithms usually employ a strongly supervised, resource-rich approach difficult to use for poorly-resourced languages. However, any approach needs to have evaluation data to confirm the results. In order to simplify the evaluation process for poorly-resourced languages (in terms of STS evaluation datasets), we present new datasets for cross-lingual and monolingual STS for languages without this evaluation data. We also present the results of several state-of-the-art methods on these data which can be used as a baseline for further research. We believe that this article will not only extend the current STS research to other languages, but will also encourage competition on this new evaluation data."
}
Markdown (Informal)
[Evaluation Datasets for Cross-lingual Semantic Textual Similarity](https://preview.aclanthology.org/jlcl-multiple-ingestion/2021.ranlp-1.59/) (Hercig & Kral, RANLP 2021)
ACL