AraBERT Model for Propaganda Detection
Mohamad Sharara, Wissam Mohamad, Ralph Tawil, Ralph Chobok, Wolf Assi, Antonio Tannoury
Abstract
Nowadays, the rapid dissemination of data on digital platforms has resulted in the emergence of information pollution and data contamination, specifically mis-information, mal-information, dis-information, fake news, and various types of propaganda. These topics are now posing a serious threat to the online digital realm, posing numerous challenges to social media platforms and governments around the world. In this article, we propose a propaganda detection model based on the transformer-based model AraBERT, with the objective of using this framework to detect propagandistic content in the Arabic social media text scene, well with purpose of making online Arabic news and media consumption healthier and safer. Given the dataset, our results are relatively encouraging, indicating a huge potential for this line of approaches in Arabic online news text NLP.- Anthology ID:
- 2022.wanlp-1.61
- Original:
- 2022.wanlp-1.61v1
- Version 2:
- 2022.wanlp-1.61v2
- Volume:
- Proceedings of the Seventh Arabic Natural Language Processing Workshop (WANLP)
- Month:
- December
- Year:
- 2022
- Address:
- Abu Dhabi, United Arab Emirates (Hybrid)
- Editors:
- Houda Bouamor, Hend Al-Khalifa, Kareem Darwish, Owen Rambow, Fethi Bougares, Ahmed Abdelali, Nadi Tomeh, Salam Khalifa, Wajdi Zaghouani
- Venue:
- WANLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 520–523
- Language:
- URL:
- https://aclanthology.org/2022.wanlp-1.61
- DOI:
- 10.18653/v1/2022.wanlp-1.61
- Cite (ACL):
- Mohamad Sharara, Wissam Mohamad, Ralph Tawil, Ralph Chobok, Wolf Assi, and Antonio Tannoury. 2022. AraBERT Model for Propaganda Detection. In Proceedings of the Seventh Arabic Natural Language Processing Workshop (WANLP), pages 520–523, Abu Dhabi, United Arab Emirates (Hybrid). Association for Computational Linguistics.
- Cite (Informal):
- AraBERT Model for Propaganda Detection (Sharara et al., WANLP 2022)
- PDF:
- https://preview.aclanthology.org/ingest-2024-clasp/2022.wanlp-1.61.pdf