Abdulhamid Abubakar


2025

pdf bib
HausaNLP at SemEval-2025 Task 2: Entity-Aware Fine-tuning vs. Prompt Engineering in Entity-Aware Machine Translation
Abdulhamid Abubakar | Hamidatu Abdulkadir | Rabiu Ibrahim | Abubakar Auwal | Ahmad Wali | Amina Umar | Maryam Bala | Sani Abdullahi Sani | Ibrahim Said Ahmad | Shamsuddeen Hassan Muhammad | Idris Abdulmumin | Vukosi Marivate
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)

This paper presents our findings for SemEval 2025 Task 2, a shared task on entity-aware machine translation (EA-MT). The goal of this task is to develop translation models that can accurately translate English sentences into target languages, with a particular focus on handling named entities, which often pose challenges for MT systems. The task covers 10 target languages with English as the source. In this paper, we describe the different systems we employed, detail our results, and discuss insights gained from our experiments.

pdf bib
HausaNLP at SemEval-2025 Task 3: Towards a Fine-Grained Model-Aware Hallucination Detection
Maryam Bala | Amina Abubakar | Abdulhamid Abubakar | Abdulkadir Bichi | Hafsa Ahmad | Sani Abdullahi Sani | Idris Abdulmumin | Shamsuddeen Hassan Muhammad | Ibrahim Said Ahmad
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)

This paper presents our findings of the Multilingual Shared Task on Hallucinations and Related Observable Overgeneration Mistakes, MU-SHROOM, which focuses on identifying hallucinations and related overgeneration errors in large language models (LLMs). The shared task involves detecting specific text spans that constitute hallucinations in the outputs generated by LLMs in 14 languages. To address this task, we aim to provide a nuanced, model-aware understanding of hallucination occurrences and severity in English. We used natural language inference and fine-tuned a ModernBERT model using a synthetic dataset of 400 samples, achieving an Intersection over Union (IoU) score of 0.032 and a correlation score of 0.422. These results indicate a moderately positive correlation between the model’s confidence scores and the actual presence of hallucinations. The IoU score indicates that our modelhas a relatively low overlap between the predicted hallucination span and the truth annotation. The performance is unsurprising, given the intricate nature of hallucination detection. Hallucinations often manifest subtly, relying on context, making pinpointing their exact boundaries formidable.

pdf bib
HausaNLP at SemEval-2025 Task 11: Advancing Hausa Text-based Emotion Detection
Sani Abdullahi Sani | Salim Abubakar | Falalu Ibrahim Lawan | Abdulhamid Abubakar | Maryam Bala
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)

This paper presents our approach to multi-label emotion detection in Hausa, a low-resource African language, as part of SemEval Track A. We fine-tuned AfriBERTa, a transformer-based model pre-trained on African languages, to classify Hausa text into six emotions: anger, disgust, fear, joy, sadness, and surprise. Our methodology involved data preprocessing, tokenization, and model fine-tuning using the Hugging Face Trainer API. The system achieved a validation accuracy of 74.00%, with an F1-score of 73.50%, demonstrating the effectiveness of transformer-based models for emotion detection in low-resource languages.