BrainT at IEST 2018: Fine-tuning Multiclass Perceptron For Implicit Emotion Classification

Vachagan Gratian, Marina Haid


Abstract
We present BrainT, a multi-class, averaged perceptron tested on implicit emotion prediction of tweets. We show that the dataset is linearly separable and explore ways in fine-tuning the baseline classifier. Our results indicate that the bag-of-words features benefit the model moderately and prediction can be improved with bigrams, trigrams, skip-one-tetragrams and POS-tags. Furthermore, we find preprocessing of the n-grams, including stemming, lowercasing, stopword filtering, emoji and emoticon conversion generally not useful. The model is trained on an annotated corpus of 153,383 tweets and predictions on the test data were submitted to the WASSA-2018 Implicit Emotion Shared Task. BrainT attained a Macro F-score of 0.63.
Anthology ID:
W18-6235
Volume:
Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis
Month:
October
Year:
2018
Address:
Brussels, Belgium
Editors:
Alexandra Balahur, Saif M. Mohammad, Veronique Hoste, Roman Klinger
Venue:
WASSA
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
243–247
Language:
URL:
https://preview.aclanthology.org/build-pipeline-with-new-library/W18-6235/
DOI:
10.18653/v1/W18-6235
Bibkey:
Cite (ACL):
Vachagan Gratian and Marina Haid. 2018. BrainT at IEST 2018: Fine-tuning Multiclass Perceptron For Implicit Emotion Classification. In Proceedings of the 9th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, pages 243–247, Brussels, Belgium. Association for Computational Linguistics.
Cite (Informal):
BrainT at IEST 2018: Fine-tuning Multiclass Perceptron For Implicit Emotion Classification (Gratian & Haid, WASSA 2018)
Copy Citation:
PDF:
https://preview.aclanthology.org/build-pipeline-with-new-library/W18-6235.pdf
Code
 ims-teamlab2018/Braint