@inproceedings{nigusie-2025-supervised,
title = "Supervised Machine Learning based {A}mharic Text Complexity Classification Using Automatic Annotator Tool",
author = "Nigusie, Gebregziabihier",
editor = "Lignos, Constantine and
Abdulmumin, Idris and
Adelani, David",
booktitle = "Proceedings of the Sixth Workshop on African Natural Language Processing (AfricaNLP 2025)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.africanlp-1.2/",
pages = "7--14",
ISBN = "979-8-89176-257-2",
abstract = "Understanding written content can vary significantly based on the linguistic complexity of the text. In the context of Amharic, a morphologically rich and low-resource language, the use of complex vocabulary and less frequent expressions often hinders understanding, particularly among readers with limited literacy skills. Such complexity poses challenges for both human comprehension and NLP applications. Addressing this complexity in Amharic is therefore important for text readability and accessibility. In this study, we developed a text complexity annotation tool using curated list of 1,113 complex Amharic terms. Utilizing this tool, we collected and annotated a dataset comprising 20,000 sentences. Based on the annotated corpus, we developed a text complexity classification model using both traditional and deep learning approaches. For traditional machine learning models, the dataset was vectorized using the Bag-of-Words representation. For deep learning and pre-trained models, we implemented embedding layers based on Word2Vec and BERT, trained on a vocabulary consisting of 24,148 tokens. The experiment is conducted using Support Vector Machine and Random Forest for classical machine learning, and Long Short-Term Memory, Bidirectional LSTM, and BERT for deep learning and pre-trained models. The classification accuracies achieved were 83.5{\%} for SVM, 80.3{\%} for RF, 84.1{\%} for LSTM, 85.0{\%} for BiLSTM, and 89.4{\%} for the BERT-based model. Among these, the BERT-based approaches shows optimal performance for text complexity classifications which have abilityto capture long-range dependencies and contextual relationships within the text."
}
Markdown (Informal)
[Supervised Machine Learning based Amharic Text Complexity Classification Using Automatic Annotator Tool](https://preview.aclanthology.org/display_plenaries/2025.africanlp-1.2/) (Nigusie, AfricaNLP 2025)
ACL