SSSD: Leveraging Pre-trained Models and Semantic Search for Semi-supervised Stance Detection

André de Sousa, Karin Becker


Abstract
Pre-trained models (PTMs) based on the Transformers architecture are trained on massive amounts of data and can capture nuances and complexities in linguistic expressions, making them a powerful tool for many natural language processing tasks. In this paper, we present SSSD (Semantic Similarity Stance Detection), a semi-supervised method for stance detection on Twitter that automatically labels a large, domain-related corpus for training a stance classification model. The method assumes as input a domain set of tweets about a given target and a labeled query set of tweets of representative arguments related to the stances. It scales the automatic labeling of a large number of tweets, and improves classification accuracy by leveraging the power of PTMs and semantic search to capture context and meaning. We largely outperformed all baselines in experiments using the Semeval benchmark.
Anthology ID:
2023.ranlp-1.30
Volume:
Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing
Month:
September
Year:
2023
Address:
Varna, Bulgaria
Editors:
Ruslan Mitkov, Galia Angelova
Venue:
RANLP
SIG:
Publisher:
INCOMA Ltd., Shoumen, Bulgaria
Note:
Pages:
264–273
Language:
URL:
https://aclanthology.org/2023.ranlp-1.30
DOI:
Bibkey:
Cite (ACL):
André de Sousa and Karin Becker. 2023. SSSD: Leveraging Pre-trained Models and Semantic Search for Semi-supervised Stance Detection. In Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing, pages 264–273, Varna, Bulgaria. INCOMA Ltd., Shoumen, Bulgaria.
Cite (Informal):
SSSD: Leveraging Pre-trained Models and Semantic Search for Semi-supervised Stance Detection (de Sousa & Becker, RANLP 2023)
Copy Citation:
PDF:
https://preview.aclanthology.org/naacl24-info/2023.ranlp-1.30.pdf