Chenran Cai
2023
MMSD2.0: Towards a Reliable Multi-modal Sarcasm Detection System
Libo Qin
|
Shijue Huang
|
Qiguang Chen
|
Chenran Cai
|
Yudi Zhang
|
Bin Liang
|
Wanxiang Che
|
Ruifeng Xu
Findings of the Association for Computational Linguistics: ACL 2023
Multi-modal sarcasm detection has attracted much recent attention. Nevertheless, the existing benchmark (MMSD) has some shortcomings that hinder the development of reliable multi-modal sarcasm detection system: (1) There are some spurious cues in MMSD, leading to the model bias learning; (2) The negative samples in MMSD are not always reasonable. To solve the aforementioned issues, we introduce MMSD2.0, a correction dataset that fixes the shortcomings of MMSD, by removing the spurious cues and re-annotating the unreasonable samples. Meanwhile, we present a novel framework called multi-view CLIP that is capable of leveraging multi-grained cues from multiple perspectives (i.e., text, image, and text-image interaction view) for multi-modal sarcasm detection. Extensive experiments show that MMSD2.0 is a valuable benchmark for building reliable multi-modal sarcasm detection systems and multi-view CLIP can significantly outperform the previous best baselines.
In-context Learning for Few-shot Multimodal Named Entity Recognition
Chenran Cai
|
Qianlong Wang
|
Bin Liang
|
Bing Qin
|
Min Yang
|
Kam-Fai Wong
|
Ruifeng Xu
Findings of the Association for Computational Linguistics: EMNLP 2023
Thanks in part to the availability of copious annotated resources for some entity categories, existing studies have achieved superior performance in multimodal named entity recognition (MNER). However, in the real-world scenario, it is infeasible to enumerate all entity categories in advance. Therefore, in this paper, we formulate a new few-shot multimodal named entity recognition (FewMNER) task, which aims to effectively locate and identify named entities for a text-image pair only using a small number of labeled examples. Further, we explore the merit of in-context learning (ICL) and propose a novel framework to deal with FewMNER, where three points are taken into account: i.e., converting visual modality, selecting useful examples, and designing an effective task demonstration. Specifically, we first employ an image caption model to convert images into textual descriptions, enabling large language models to absorb information from visual modality. Then, we use the ranking of the sum of similarity rankings from both text and image modalities to select k-nearest examples, which form a demonstration context. Finally, we utilize the MNER definition and the meaning of each entity category as effective instruction. Extensive experimental results demonstrate that our framework outperforms baselines under several few-shot settings.
Search
Co-authors
- Bin Liang 2
- Ruifeng Xu 2
- Libo Qin 1
- Shijue Huang 1
- Qiguang Chen 1
- show all...