Jeonghun Yeo
2024
Let’s Go Real Talk: Spoken Dialogue Model for Face-to-Face Conversation
Se Park
|
Chae Kim
|
Hyeongseop Rha
|
Minsu Kim
|
Joanna Hong
|
Jeonghun Yeo
|
Yong Ro
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
In this paper, we introduce a novel Face-to-Face spoken dialogue model. It processes audio-visual speech from user input and generates audio-visual speech as the response, marking the initial step towards creating an avatar chatbot system without relying on intermediate text. To this end, we newly introduce MultiDialog, the first large-scale multimodal (i.e, audio and visual) spoken dialogue corpus containing 340 hours of approximately 9,000 dialogues, recorded based on the open domain dialogue dataset, TopicalChat. The MultiDialog contains parallel audio-visual recordings of conversation partners acting according to the given script with emotion annotations, which we expect to open up research opportunities in multimodal synthesis. Our Face-to-Face spoken dialogue model incorporates a textually pretrained large language model and adapts it into the audio-visual spoken dialogue domain by incorporating speech-text joint pretraining. Through extensive experiments, we validate the effectiveness of our model in facilitating a face-to-face conversation. Demo and data are available at https://multidialog.github.io and https://huggingface.co/datasets/IVLLab/MultiDialog, respectively.
Search
Co-authors
- Se Park 1
- Chae Kim 1
- Hyeongseop Rha 1
- Minsu Kim 1
- Joanna Hong 1
- show all...
- Yong Ro 1
Venues
- acl1