Back Transcription as a Method for Evaluating Robustness of Natural Language Understanding Models to Speech Recognition Errors
Marek Kubis, Paweł Skórzewski, Marcin Sowański, Tomasz Zietkiewicz
Abstract
In a spoken dialogue system, an NLU model is preceded by a speech recognition system that can deteriorate the performance of natural language understanding. This paper proposes a method for investigating the impact of speech recognition errors on the performance of natural language understanding models. The proposed method combines the back transcription procedure with a fine-grained technique for categorizing the errors that affect the performance of NLU models. The method relies on the usage of synthesized speech for NLU evaluation. We show that the use of synthesized speech in place of audio recording does not change the outcomes of the presented technique in a significant way.- Anthology ID:
- 2023.emnlp-main.724
- Volume:
- Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing
- Month:
- December
- Year:
- 2023
- Address:
- Singapore
- Editors:
- Houda Bouamor, Juan Pino, Kalika Bali
- Venue:
- EMNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 11824–11835
- Language:
- URL:
- https://aclanthology.org/2023.emnlp-main.724
- DOI:
- 10.18653/v1/2023.emnlp-main.724
- Cite (ACL):
- Marek Kubis, Paweł Skórzewski, Marcin Sowański, and Tomasz Zietkiewicz. 2023. Back Transcription as a Method for Evaluating Robustness of Natural Language Understanding Models to Speech Recognition Errors. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 11824–11835, Singapore. Association for Computational Linguistics.
- Cite (Informal):
- Back Transcription as a Method for Evaluating Robustness of Natural Language Understanding Models to Speech Recognition Errors (Kubis et al., EMNLP 2023)
- PDF:
- https://preview.aclanthology.org/naacl24-info/2023.emnlp-main.724.pdf