@inproceedings{zhang-etal-2023-cross-lingual,
title = "Cross-Lingual Cross-Target Stance Detection with Dual Knowledge Distillation Framework",
author = "Zhang, Ruike and
Yang, Hanxuan and
Mao, Wenji",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.666/",
doi = "10.18653/v1/2023.emnlp-main.666",
pages = "10804--10819",
abstract = "Stance detection aims to identify the user{'}s attitude toward specific \textit{targets} from text, which is an important research area in text mining and benefits a variety of application domains. Existing studies on stance detection were conducted mainly in English. Due to the low-resource problem in most non-English languages, cross-lingual stance detection was proposed to transfer knowledge from high-resource (source) language to low-resource (target) language. However, previous research has ignored the practical issue of no labeled training data available in target language. Moreover, target inconsistency in cross-lingual stance detection brings about the additional issue of unseen targets in target language, which in essence requires the transfer of both language and target-oriented knowledge from source to target language. To tackle these challenging issues, in this paper, we propose the new task of cross-lingual cross-target stance detection and develop the first computational work with dual knowledge distillation. Our proposed framework designs a cross-lingual teacher and a cross-target teacher using the source language data and a dual distillation process that transfers the two types of knowledge to target language. To bridge the target discrepancy between languages, cross-target teacher mines target category information and generalizes it to the unseen targets in target language via category-oriented learning. Experimental results on multilingual stance datasets demonstrate the effectiveness of our method compared to the competitive baselines."
}
Markdown (Informal)
[Cross-Lingual Cross-Target Stance Detection with Dual Knowledge Distillation Framework](https://preview.aclanthology.org/fix-sig-urls/2023.emnlp-main.666/) (Zhang et al., EMNLP 2023)
ACL