Abstract
This position paper presents my research interest in establishing human-like chat-oriented dialogue systems. To this end, my work focuses on two main areas: the construction and utilization of multimodal datasets and real-time multimodal affective computing. I discuss the limitations of current multimodal dialogue corpora and multimodal affective computing models. As a solution, I have constructed a human-human dialogue dataset containing various synchronized multimodal information, and I have conducted preliminary analyses on it. In future work, I will further analyze the collected data and build a real-time multimodal emotion estimation model for dialogue systems.- Anthology ID:
- 2024.yrrsds-1.22
- Volume:
- Proceedings of the 20th Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems
- Month:
- September
- Year:
- 2024
- Address:
- Kyoto, Japan
- Editors:
- Koji Inoue, Yahui Fu, Agnes Axelsson, Atsumoto Ohashi, Brielen Madureira, Yuki Zenimoto, Biswesh Mohapatra, Armand Stricker, Sopan Khosla
- Venues:
- YRRSDS | WS
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 60–61
- Language:
- URL:
- https://aclanthology.org/2024.yrrsds-1.22
- DOI:
- Cite (ACL):
- Jingjing Jiang. 2024. Towards a Real-Time Multimodal Emotion Estimation Model for Dialogue Systems. In Proceedings of the 20th Workshop of Young Researchers' Roundtable on Spoken Dialogue Systems, pages 60–61, Kyoto, Japan. Association for Computational Linguistics.
- Cite (Informal):
- Towards a Real-Time Multimodal Emotion Estimation Model for Dialogue Systems (Jiang, YRRSDS-WS 2024)
- PDF:
- https://preview.aclanthology.org/dois-2013-emnlp/2024.yrrsds-1.22.pdf