Abstract
This paper describes the system of the LowResource Team for Task 2 of BLP-2023, which involves conducting sentiment analysis on a dataset composed of public posts and comments from diverse social media platforms. Our primary aim was to utilize BanglaBert, a BERT model pre-trained on a large Bangla corpus, using various strategies including fine-tuning, dropping random tokens, and using several external datasets. Our final model is an ensemble of the three best BanglaBert variations. Our system achieved overall 3rd in the Test Set among 30 participating teams with a score of 0.718. Additionally, we discuss the promising systems that didn’t perform well namely task-adaptive pertaining and paraphrasing using BanglaT5. Our training codes are publicly available at https://github.com/Aunabil4602/bnlp-workshop-task2-2023- Anthology ID:
- 2023.banglalp-1.47
- Volume:
- Proceedings of the First Workshop on Bangla Language Processing (BLP-2023)
- Month:
- December
- Year:
- 2023
- Address:
- Singapore
- Editors:
- Firoj Alam, Sudipta Kar, Shammur Absar Chowdhury, Farig Sadeque, Ruhul Amin
- Venue:
- BanglaLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 347–353
- Language:
- URL:
- https://aclanthology.org/2023.banglalp-1.47
- DOI:
- 10.18653/v1/2023.banglalp-1.47
- Cite (ACL):
- Aunabil Chakma and Masum Hasan. 2023. LowResource at BLP-2023 Task 2: Leveraging BanglaBert for Low Resource Sentiment Analysis of Bangla Language. In Proceedings of the First Workshop on Bangla Language Processing (BLP-2023), pages 347–353, Singapore. Association for Computational Linguistics.
- Cite (Informal):
- LowResource at BLP-2023 Task 2: Leveraging BanglaBert for Low Resource Sentiment Analysis of Bangla Language (Chakma & Hasan, BanglaLP 2023)
- PDF:
- https://preview.aclanthology.org/landing_page/2023.banglalp-1.47.pdf