@inproceedings{xu-etal-2022-mmdag,
title = "{MMDAG}: Multimodal Directed Acyclic Graph Network for Emotion Recognition in Conversation",
author = "Xu, Shuo and
Jia, Yuxiang and
Niu, Changyong and
Zan, Hongying",
editor = "Calzolari, Nicoletta and
B{\'e}chet, Fr{\'e}d{\'e}ric and
Blache, Philippe and
Choukri, Khalid and
Cieri, Christopher and
Declerck, Thierry and
Goggi, Sara and
Isahara, Hitoshi and
Maegaard, Bente and
Mariani, Joseph and
Mazo, H{\'e}l{\`e}ne and
Odijk, Jan and
Piperidis, Stelios",
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
month = jun,
year = "2022",
address = "Marseille, France",
publisher = "European Language Resources Association",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.lrec-1.733/",
pages = "6802--6807",
abstract = "Emotion recognition in conversation is important for an empathetic dialogue system to understand the user{'}s emotion and then generate appropriate emotional responses. However, most previous researches focus on modeling conversational contexts primarily based on the textual modality or simply utilizing multimodal information through feature concatenation. In order to exploit multimodal information and contextual information more effectively, we propose a multimodal directed acyclic graph (MMDAG) network by injecting information flows inside modality and across modalities into the DAG architecture. Experiments on IEMOCAP and MELD show that our model outperforms other state-of-the-art models. Comparative studies validate the effectiveness of the proposed modality fusion method."
}
Markdown (Informal)
[MMDAG: Multimodal Directed Acyclic Graph Network for Emotion Recognition in Conversation](https://preview.aclanthology.org/fix-sig-urls/2022.lrec-1.733/) (Xu et al., LREC 2022)
ACL