@inproceedings{acharya-etal-2025-paramananda,
title = "Paramananda@{NLU} of {D}evanagari Script Languages 2025: Detection of Language, Hate Speech and Targets using {F}ast{T}ext and {BERT}",
author = "Acharya, Darwin and
Dawadi, Sundeep and
Saud, Shivram and
Regmi, Sunil",
editor = "Sarveswaran, Kengatharaiyer and
Vaidya, Ashwini and
Krishna Bal, Bal and
Shams, Sana and
Thapa, Surendrabikram",
booktitle = "Proceedings of the First Workshop on Challenges in Processing South Asian Languages (CHiPSAL 2025)",
month = jan,
year = "2025",
address = "Abu Dhabi, UAE",
publisher = "International Committee on Computational Linguistics",
url = "https://preview.aclanthology.org/ingest_wac_2008/2025.chipsal-1.39/",
pages = "334--338",
abstract = "This paper presents a comparative analysis of FastText and BERT-based approaches for Natural Language Understanding (NLU) tasks in Devanagari script languages. We evaluate these models on three critical tasks: language identification, hate speech detection, and target identification across five languages: Nepali, Marathi, Sanskrit, Bhojpuri, and Hindi. Our experiments, although with raw tweet dataset but extracting only devanagari script, demonstrate that while both models achieve exceptional performance in language identification (F1 scores {\ensuremath{>}} 0.99), they show varying effectiveness in hate speech detection and target identification tasks. FastText with augmented data outperforms BERT in hate speech detection (F1 score: 0.8552 vs 0.5763), while BERT shows superior performance in target identification (F1 score: 0.5785 vs 0.4898). These findings contribute to the growing body of research on NLU for low-resource languages and provide insights into model selection for specific tasks in Devanagari script processing."
}
Markdown (Informal)
[Paramananda@NLU of Devanagari Script Languages 2025: Detection of Language, Hate Speech and Targets using FastText and BERT](https://preview.aclanthology.org/ingest_wac_2008/2025.chipsal-1.39/) (Acharya et al., CHiPSAL 2025)
ACL