Sedigh Khademi
2024
Enhancing Social Media Health Prediction Certainty by Integrating Large Language Models with Transformer Classifiers
Sedigh Khademi
|
Christopher Palmer
|
Muhammad Javed
|
Jim Buttery
|
Gerardo Dimaguila
Proceedings of The 9th Social Media Mining for Health Research and Applications (SMM4H 2024) Workshop and Shared Tasks
This paper presents our approach for SMM4H 2024 Task 5, focusing on identifying tweets where users discuss their child’s health conditions of ADHD, ASD, delayed speech, or asthma. Our approach uses a pipeline that combines transformer-based classifiers and GPT-4 large language models (LLMs). We first address data imbalance in the training set using topic modelling and under-sampling. Next, we train RoBERTa-based classifiers on the adjusted data. Finally, GPT-4 refines the classifier’s predictions for uncertain cases (confidence below 0.9). This strategy achieved significant improvement over the baseline RoBERTa models. Our work demonstrates the effectiveness of combining transformer classifiers and LLMs for extracting health insights from social media conversations.
2020
Adverse Drug Reaction Detection in Twitter Using RoBERTa and Rules
Sedigh Khademi
|
Pari Delirhaghighi
|
Frada Burstein
Proceedings of the Fifth Social Media Mining for Health Applications Workshop & Shared Task
This paper describes the method we developed for the Task 2 English variation of the Social Media Mining for Health Applications (SMM4H) 2020 shared task. The task was to classify tweets containing adverse effects (AE) after medication intake. Our approach combined transfer learning using a RoBERTa Large Transformer model with a rule-based post-prediction correction to improve model precision. The model’s F1-Score of 0.56 on the test dataset was 10% better than the mean of the F1-Score of the best submissions in the task.
Search
Co-authors
- Christopher Palmer 1
- Muhammad Javed 1
- Jim Buttery 1
- Gerardo Dimaguila 1
- Pari Delirhaghighi 1
- show all...