Abstract
Recently, neural twitter sentiment classification has become one of state-of-thearts, which relies less feature engineering work compared with traditional methods. In this paper, we propose a simple and effective ensemble method to further boost the performances of neural models. We collect several word embedding sets which are publicly released (often are learned on different corpus) or constructed by running Skip-gram on released large-scale corpus. We make an assumption that different word embeddings cover different words and encode different semantic knowledge, thus using them together can improve the generalizations and performances of neural models. In the SemEval 2017, our method ranks 1st in Accuracy, 5th in AverageR. Meanwhile, the additional comparisons demonstrate the superiority of our model over these ones based on only one word embedding set. We release our code for the method duplicability.- Anthology ID:
- S17-2102
- Volume:
- Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017)
- Month:
- August
- Year:
- 2017
- Address:
- Vancouver, Canada
- Editors:
- Steven Bethard, Marine Carpuat, Marianna Apidianaki, Saif M. Mohammad, Daniel Cer, David Jurgens
- Venue:
- SemEval
- SIG:
- SIGLEX
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 621–625
- Language:
- URL:
- https://aclanthology.org/S17-2102
- DOI:
- 10.18653/v1/S17-2102
- Cite (ACL):
- Yichun Yin, Yangqiu Song, and Ming Zhang. 2017. NNEMBs at SemEval-2017 Task 4: Neural Twitter Sentiment Classification: a Simple Ensemble Method with Different Embeddings. In Proceedings of the 11th International Workshop on Semantic Evaluation (SemEval-2017), pages 621–625, Vancouver, Canada. Association for Computational Linguistics.
- Cite (Informal):
- NNEMBs at SemEval-2017 Task 4: Neural Twitter Sentiment Classification: a Simple Ensemble Method with Different Embeddings (Yin et al., SemEval 2017)
- PDF:
- https://preview.aclanthology.org/naacl-24-ws-corrections/S17-2102.pdf