Abstract
Pre-trained models (PTMs) based on the Transformers architecture are trained on massive amounts of data and can capture nuances and complexities in linguistic expressions, making them a powerful tool for many natural language processing tasks. In this paper, we present SSSD (Semantic Similarity Stance Detection), a semi-supervised method for stance detection on Twitter that automatically labels a large, domain-related corpus for training a stance classification model. The method assumes as input a domain set of tweets about a given target and a labeled query set of tweets of representative arguments related to the stances. It scales the automatic labeling of a large number of tweets, and improves classification accuracy by leveraging the power of PTMs and semantic search to capture context and meaning. We largely outperformed all baselines in experiments using the Semeval benchmark.- Anthology ID:
- 2023.ranlp-1.30
- Volume:
- Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing
- Month:
- September
- Year:
- 2023
- Address:
- Varna, Bulgaria
- Editors:
- Ruslan Mitkov, Galia Angelova
- Venue:
- RANLP
- SIG:
- Publisher:
- INCOMA Ltd., Shoumen, Bulgaria
- Note:
- Pages:
- 264–273
- Language:
- URL:
- https://aclanthology.org/2023.ranlp-1.30
- DOI:
- Cite (ACL):
- André de Sousa and Karin Becker. 2023. SSSD: Leveraging Pre-trained Models and Semantic Search for Semi-supervised Stance Detection. In Proceedings of the 14th International Conference on Recent Advances in Natural Language Processing, pages 264–273, Varna, Bulgaria. INCOMA Ltd., Shoumen, Bulgaria.
- Cite (Informal):
- SSSD: Leveraging Pre-trained Models and Semantic Search for Semi-supervised Stance Detection (de Sousa & Becker, RANLP 2023)
- PDF:
- https://preview.aclanthology.org/naacl24-info/2023.ranlp-1.30.pdf