@inproceedings{alkanhal-etal-2023-aswat,
    title = "Aswat: {A}rabic Audio Dataset for Automatic Speech Recognition Using Speech-Representation Learning",
    author = "Alkanhal, Lamya  and
      Alessa, Abeer  and
      Almahmoud, Elaf  and
      Alaqil, Rana",
    editor = "Sawaf, Hassan  and
      El-Beltagy, Samhaa  and
      Zaghouani, Wajdi  and
      Magdy, Walid  and
      Abdelali, Ahmed  and
      Tomeh, Nadi  and
      Abu Farha, Ibrahim  and
      Habash, Nizar  and
      Khalifa, Salam  and
      Keleg, Amr  and
      Haddad, Hatem  and
      Zitouni, Imed  and
      Mrini, Khalil  and
      Almatham, Rawan",
    booktitle = "Proceedings of ArabicNLP 2023",
    month = dec,
    year = "2023",
    address = "Singapore (Hybrid)",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2023.arabicnlp-1.10/",
    doi = "10.18653/v1/2023.arabicnlp-1.10",
    pages = "120--127",
    abstract = "Recent advancements in self-supervised speech-representation learning for automatic speech recognition (ASR) approaches have significantly improved the results on many benchmarks with low-cost data labeling. In this paper, we train two self-supervised frameworks for ASR, namely wav2vec, and data2vec, in which we conduct multiple experiments and analyze their results. Furthermore, we introduce Aswat dataset, which covers multiple genres and features speakers with vocal variety. Aswat contains 732 hours of clean Arabic speech that can be used in the pretraining task for learning latent speech representations, which results in achieving a lower word error rate (WER) in Arabic ASR. We report the baseline results and achieve state-of-the-art WERs of 11.7{\%} and 10.3{\%} on Common Voice (CV) and the second round of Multi-Genre Broadcast (MGB-2) respectively, as a result of including our dataset Aswat."
}Markdown (Informal)
[Aswat: Arabic Audio Dataset for Automatic Speech Recognition Using Speech-Representation Learning](https://preview.aclanthology.org/ingest-emnlp/2023.arabicnlp-1.10/) (Alkanhal et al., ArabicNLP 2023)
ACL