@inproceedings{xu-etal-2023-mpmr,
title = "m{PMR}: A Multilingual Pre-trained Machine Reader at Scale",
author = "Xu, Weiwen and
Li, Xin and
Lam, Wai and
Bing, Lidong",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest_wac_2008/2023.acl-short.131/",
doi = "10.18653/v1/2023.acl-short.131",
pages = "1533--1546",
abstract = "We present multilingual Pre-trained Machine Reader (mPMR), a novel method for multilingual machine reading comprehension (MRC)-style pre-training. mPMR aims to guide multilingual pre-trained language models (mPLMs) to perform natural language understanding (NLU) including both sequence classification and span extraction in multiple languages. To achieve cross-lingual generalization when only source-language fine-tuning data is available, existing mPLMs solely transfer NLU capability from a source language to target languages. In contrast, mPMR allows the direct inheritance of multilingual NLU capability from the MRC-style pre-training to downstream tasks. Therefore, mPMR acquires better NLU capability for target languages. mPMR also provides a unified solver for tackling cross-lingual span extraction and sequence classification, thereby enabling the extraction of rationales to explain the sentence-pair classification process."
}
Markdown (Informal)
[mPMR: A Multilingual Pre-trained Machine Reader at Scale](https://preview.aclanthology.org/ingest_wac_2008/2023.acl-short.131/) (Xu et al., ACL 2023)
ACL
- Weiwen Xu, Xin Li, Wai Lam, and Lidong Bing. 2023. mPMR: A Multilingual Pre-trained Machine Reader at Scale. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 1533–1546, Toronto, Canada. Association for Computational Linguistics.