@inproceedings{briakou-carpuat-2020-detecting,
    title = "{D}etecting {F}ine-{G}rained {C}ross-{L}ingual {S}emantic {D}ivergences without {S}upervision by {L}earning to {R}ank",
    author = "Briakou, Eleftheria  and
      Carpuat, Marine",
    editor = "Webber, Bonnie  and
      Cohn, Trevor  and
      He, Yulan  and
      Liu, Yang",
    booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
    month = nov,
    year = "2020",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2020.emnlp-main.121/",
    doi = "10.18653/v1/2020.emnlp-main.121",
    pages = "1563--1580",
    abstract = "Detecting fine-grained differences in content conveyed in different languages matters for cross-lingual NLP and multilingual corpora analysis, but it is a challenging machine learning problem since annotation is expensive and hard to scale. This work improves the prediction and annotation of fine-grained semantic divergences. We introduce a training strategy for multilingual BERT models by learning to rank synthetic divergent examples of varying granularity. We evaluate our models on the Rationalized English-French Semantic Divergences, a new dataset released with this work, consisting of English-French sentence-pairs annotated with semantic divergence classes and token-level rationales. Learning to rank helps detect fine-grained sentence-level divergences more accurately than a strong sentence-level similarity model, while token-level predictions have the potential of further distinguishing between coarse and fine-grained divergences."
}Markdown (Informal)
[Detecting Fine-Grained Cross-Lingual Semantic Divergences without Supervision by Learning to Rank](https://preview.aclanthology.org/ingest-emnlp/2020.emnlp-main.121/) (Briakou & Carpuat, EMNLP 2020)
ACL