@inproceedings{yadav-sitaram-2022-survey,
title = "A Survey of Multilingual Models for Automatic Speech Recognition",
author = "Yadav, Hemant and
Sitaram, Sunayana",
editor = "Calzolari, Nicoletta and
B{\'e}chet, Fr{\'e}d{\'e}ric and
Blache, Philippe and
Choukri, Khalid and
Cieri, Christopher and
Declerck, Thierry and
Goggi, Sara and
Isahara, Hitoshi and
Maegaard, Bente and
Mariani, Joseph and
Mazo, H{\'e}l{\`e}ne and
Odijk, Jan and
Piperidis, Stelios",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2022.lrec-1.542/",
pages = "5071--5079",
abstract = "Although Automatic Speech Recognition (ASR) systems have achieved human-like performance for a few languages, the majority of the world`s languages do not have usable systems due to the lack of large speech datasets to train these models. Cross-lingual transfer is an attractive solution to this problem, because low-resource languages can potentially benefit from higher-resource languages either through transfer learning, or being jointly trained in the same multilingual model. The problem of cross-lingual transfer has been well studied in ASR, however, recent advances in Self Supervised Learning are opening up avenues for unlabeled speech data to be used in multilingual ASR models, which can pave the way for improved performance on low-resource languages. In this paper, we survey the state of the art in multilingual ASR models that are built with cross-lingual transfer in mind. We present best practices for building multilingual models from research across diverse languages and techniques, discuss open questions and provide recommendations for future work."
}
Markdown (Informal)
[A Survey of Multilingual Models for Automatic Speech Recognition](https://preview.aclanthology.org/add-emnlp-2024-awards/2022.lrec-1.542/) (Yadav & Sitaram, LREC 2022)
ACL