Paralinguistic Attitude Recognition for Spoken Dialogue Systems

Kouki Miyazawa, Zhi Zhu, Yoshinao Sato


Abstract
Although paralinguistic information is critical for human communication, most spoken dialogue systems ignore such information, hindering natural communication between humans and machines. This study addresses the recognition of paralinguistic attitudes in user speech. Specifically, we focus on four essential attitudes for generating an appropriate system response, namely agreement, disagreement, questions, and stalling. The proposed model can help a dialogue system better understand what the user is trying to convey. In our experiments, we trained and evaluated a model that classified paralinguistic attitudes on a reading-speech dataset without using linguistic information. The proposed model outperformed human perception. Furthermore, experimental results indicate that speech enhancement alleviates the degradation of model performance caused by background noise, whereas reverberation remains a challenge.
Anthology ID:
2025.iwsds-1.11
Volume:
Proceedings of the 15th International Workshop on Spoken Dialogue Systems Technology
Month:
May
Year:
2025
Address:
Bilbao, Spain
Editors:
Maria Ines Torres, Yuki Matsuda, Zoraida Callejas, Arantza del Pozo, Luis Fernando D'Haro
Venues:
IWSDS | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
137–142
Language:
URL:
https://preview.aclanthology.org/fix-sig-urls/2025.iwsds-1.11/
DOI:
Bibkey:
Cite (ACL):
Kouki Miyazawa, Zhi Zhu, and Yoshinao Sato. 2025. Paralinguistic Attitude Recognition for Spoken Dialogue Systems. In Proceedings of the 15th International Workshop on Spoken Dialogue Systems Technology, pages 137–142, Bilbao, Spain. Association for Computational Linguistics.
Cite (Informal):
Paralinguistic Attitude Recognition for Spoken Dialogue Systems (Miyazawa et al., IWSDS 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/fix-sig-urls/2025.iwsds-1.11.pdf