Sukumar Nandi
2026
Sentence-Level Back-Transliteration of Romanized Indian Languages: Performance Analysis and Challenges
Saurabh Kumar | Dhruvkumar Babubhai Kakadiya | Sanasam Ranbir Singh | Sukumar Nandi
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Saurabh Kumar | Dhruvkumar Babubhai Kakadiya | Sanasam Ranbir Singh | Sukumar Nandi
Proceedings of the Fifteenth Language Resources and Evaluation Conference
The widespread use of Romanized text for Indian languages, particularly on social media platforms, poses significant challenges for natural language processing due to the lack of standardized orthography and the presence of contextual ambiguities. In this study, we explore sentence-level back-transliteration for 13 Indian languages, focusing on addressing the limitations of word-level models that fail to capture contextual dependencies. We evaluate state-of-the-art models, including fine-tuned LLaMA, mT5, and Multilingual Transformer models, comparing their performance against the baseline IndicXlit model. In addition, we conduct a comprehensive error analysis to gain deeper insights into model performance. Our results demonstrate that fine-tuned LLaMA and the proposed IndiXform model, specifically designed to leverage sentence-level context, significantly outperform zero-shot LLaMA and the IndicXlit baseline. These findings provide valuable insights into handling contextual ambiguities and enhancing the accuracy of back-transliteration systems for Indian languages.
AssamLegalTrans: A Parallel Corpus, Benchmark and Analysis for English-Assamese Machine Translation of Legal Judgments
Telem Joyson Singh | Hemanta Baruah | Sanasam Ranbir Singh | Anindita Talukdar | Nasrin Shahnaz | Okram Jimmy Singh | Priyankoo Sarmah | Pallav Kumar Dutta | Sukumar Nandi | Pranab Duara
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Telem Joyson Singh | Hemanta Baruah | Sanasam Ranbir Singh | Anindita Talukdar | Nasrin Shahnaz | Okram Jimmy Singh | Priyankoo Sarmah | Pallav Kumar Dutta | Sukumar Nandi | Pranab Duara
Proceedings of the Fifteenth Language Resources and Evaluation Conference
In India, the official language for writing judgments in higher courts is English, which creates a language barrier for citizens not proficient in English. Machine Translation (MT) provides a scalable solution, but its progress for low-resource languages like Assamese is significantly limited due to the lack of legal domain data. To address this gap, we introduce the first-of-its-kind English-Assamese parallel corpus for the translation of Indian court judgments. This dataset consists of over 55,000 manually translated and validated sentence pairs from over 500 judgments of the Gauhati High Court and the Supreme Court of India. Using this dataset, we perform a comprehensive evaluation of state-of-the-art multilingual models, including NLLB-200 and Sarvam-Translate, in both zero-shot and fine-tuned settings, comparing their performance against commercial systems. Our experiments show that fine-tuning on our legal-domain dataset significantly improves the translation quality. We also conduct a thorough error analysis that points out important issues in legal translation. These include precisely translating legal terms, properly transliterating named entities, expanding abbreviations, and transforming sentence structures, such as changing passive voice to active voice, when translating from English to Assamese. By creating a publicly available dataset and examining the specific challenges, this work offers a reproducible foundation and a clear way to develop more accurate and reliable legal machine translation systems. This will help improve access to justice for Assamese speakers.
2025
AsRED: Development and Evaluation of an Assamese Reduplication Dataset
Pankaj Choudhury | Chaitanya Kirti | Dhrubajyoti Pathak | Sukumar Nandi
Proceedings of the 39th Pacific Asia Conference on Language, Information and Computation
Pankaj Choudhury | Chaitanya Kirti | Dhrubajyoti Pathak | Sukumar Nandi
Proceedings of the 39th Pacific Asia Conference on Language, Information and Computation
indiDataMiner at SemEval-2025 Task 11: From Text to Emotion: Transformer-Based Models for Emotions Detection in Indian Languages
Saurabh Kumar | Sujit Kumar | Sanasam Ranbir Singh | Sukumar Nandi
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
Saurabh Kumar | Sujit Kumar | Sanasam Ranbir Singh | Sukumar Nandi
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
Emotion detection is essential for applications like mental health monitoring and social media analysis, yet remains underexplored for Indian languages. This paper presents our system for SemEval-2025 Task 11 (Track A), focusing on multilabel emotion detection in Hindi and Marathi, two widely spoken Indian languages. We fine-tune IndicBERT v2 on the BRIGHTER dataset, achieving F1 scores of 87.37 (Hindi) and 88.32 (Marathi), outperforming baseline models. Our results highlight the effectiveness of fine-tuning a language-specific pretrained model for emotion detection, contributing to advancements in multilingual NLP research.
2024
IndiSentiment140: Sentiment Analysis Dataset for Indian Languages with Emphasis on Low-Resource Languages using Machine Translation
Saurabh Kumar | Ranbir Sanasam | Sukumar Nandi
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Saurabh Kumar | Ranbir Sanasam | Sukumar Nandi
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Sentiment analysis, a fundamental aspect of Natural Language Processing (NLP), involves the classification of emotions, opinions, and attitudes in text data. In the context of India, with its vast linguistic diversity and low-resource languages, the challenge is to support sentiment analysis in numerous Indian languages. This study explores the use of machine translation to bridge this gap. The investigation examines the feasibility of machine translation for creating sentiment analysis datasets in 22 Indian languages. Google Translate, with its extensive language support, is employed for this purpose in translating the Sentiment140 dataset. The study aims to provide insights into the practicality of using machine translation in the context of India’s linguistic diversity for sentiment analysis datasets. Our findings indicate that a dataset generated using Google Translate has the potential to serve as a foundational framework for tackling the low-resource challenges commonly encountered in sentiment analysis for Indian languages.
Evaluating Performance of Pre-trained Word Embeddings on Assamese, a Low-resource Language
Dhrubajyoti Pathak | Sukumar Nandi | Priyankoo Sarmah
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Dhrubajyoti Pathak | Sukumar Nandi | Priyankoo Sarmah
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Word embeddings and Language models are the building blocks of modern Deep Neural Network-based Natural Language Processing. They are extensively explored in high-resource languages and provide state-of-the-art (SOTA) performance for a wide range of downstream tasks. Nevertheless, these word embeddings are not explored in languages such as Assamese, where resources are limited. Furthermore, there has been limited study into the performance evaluation of these word embeddings for low-resource languages in downstream tasks. In this research, we explore the current state of Assamese pre-trained word embeddings. We evaluate these embeddings’ performance on sequence labeling tasks such as Parts-of-speech and Named Entity Recognition. In order to assess the efficiency of the embeddings, experiments are performed utilizing both ensemble and individual word embedding approaches. The ensembling approach that uses three word embeddings outperforms the others. In the paper, the outcomes of the investigations are described. The results of this comparative performance evaluation may assist researchers in choosing an Assamese pre-trained word embedding for subsequent tasks.
2023
Image Caption Synthesis for Low Resource Assamese Language using Bi-LSTM with Bilinear Attention
Pankaj Choudhury | Prithwijit Guha | Sukumar Nandi
Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation
Pankaj Choudhury | Prithwijit Guha | Sukumar Nandi
Proceedings of the 37th Pacific Asia Conference on Language, Information and Computation
IndiSocialFT: Multilingual Word Representation for Indian languages in code-mixed environment
Saurabh Kumar | Ranbir Sanasam | Sukumar Nandi
Findings of the Association for Computational Linguistics: EMNLP 2023
Saurabh Kumar | Ranbir Sanasam | Sukumar Nandi
Findings of the Association for Computational Linguistics: EMNLP 2023
The increasing number of Indian language users on the internet necessitates the development of Indian language technologies. In response to this demand, our paper presents a generalized representation vector for diverse text characteristics, including native scripts, transliterated text, multilingual, code-mixed, and social media-related attributes. We gather text from both social media and well-formed sources and utilize the FastText model to create the “IndiSocialFT” embedding. Through intrinsic and extrinsic evaluation methods, we compare IndiSocialFT with three popular pretrained embeddings trained over Indian languages. Our findings show that the proposed embedding surpasses the baselines in most cases and languages, demonstrating its suitability for various NLP applications.
2022
AsNER - Annotated Dataset and Baseline for Assamese Named Entity recognition
Dhrubajyoti Pathak | Sukumar Nandi | Priyankoo Sarmah
Proceedings of the Thirteenth Language Resources and Evaluation Conference
Dhrubajyoti Pathak | Sukumar Nandi | Priyankoo Sarmah
Proceedings of the Thirteenth Language Resources and Evaluation Conference
We present the AsNER, a named entity annotation dataset for low resource Assamese language with a baseline Assamese NER model. The dataset contains about 99k tokens comprised of text from the speech of the Prime Minister of India and Assamese play. It also contains person names, location names and addresses. The proposed NER dataset is likely to be a significant resource for deep neural based Assamese language processing. We benchmark the dataset by training NER models and evaluating using state-of-the-art architectures for supervised named entity recognition (NER) such as Fasttext, BERT, XLM-R, FLAIR, MuRIL etc. We implement several baseline approaches with state-of-the-art sequence tagging Bi-LSTM-CRF architecture. The highest F1-score among all baselines achieves an accuracy of 80.69% when using MuRIL as a word embedding method. The annotated dataset and the top performing model are made publicly available.
Generating Monolingual Dataset for Low Resource Language Bodo from old books using Google Keep
Sanjib Narzary | Maharaj Brahma | Mwnthai Narzary | Gwmsrang Muchahary | Pranav Kumar Singh | Apurbalal Senapati | Sukumar Nandi | Bidisha Som
Proceedings of the Thirteenth Language Resources and Evaluation Conference
Sanjib Narzary | Maharaj Brahma | Mwnthai Narzary | Gwmsrang Muchahary | Pranav Kumar Singh | Apurbalal Senapati | Sukumar Nandi | Bidisha Som
Proceedings of the Thirteenth Language Resources and Evaluation Conference
Bodo is a scheduled Indian language spoken largely by the Bodo community of Assam and other northeastern Indian states. Due to a lack of resources, it is difficult for young languages to communicate more effectively with the rest of the world. This leads to a lack of research in low-resource languages. The creation of a dataset is a tedious and costly process, particularly for languages with no participatory research. This is more visible for languages that are young and have recently adopted standard writing scripts. In this paper, we present a methodology using Google Keep for OCR to generate a monolingual Bodo corpus from different books. In this work, a Bodo text corpus of 192,327 tokens and 32,268 unique tokens is generated using free, accessible, and daily-usable applications. Moreover, some essential characteristics of the Bodo language are discussed that are neglected by Natural Language Progressing (NLP) researchers.
Search
Fix author
Co-authors
- Saurabh Kumar 4
- Dhrubajyoti Pathak 3
- Priyankoo Sarmah 3
- Sanasam Ranbir Singh 3
- Pankaj Choudhury 2
- Ranbir Sanasam 2
- Hemanta Baruah 1
- Maharaj Brahma 1
- Pranab Duara 1
- Pallav Kumar Dutta 1
- Prithwijit Guha 1
- Dhruvkumar Babubhai Kakadiya 1
- Chaitanya Kirti 1
- Sujit Kumar 1
- Gwmsrang Muchahary 1
- Sanjib Narzary 1
- Mwnthai Narzary 1
- Apurbalal Senapati 1
- Nasrin Shahnaz 1
- Telem Joyson Singh 1
- Okram Jimmy Singh 1
- Pranav Kumar Singh 1
- Bidisha Som 1
- Anindita Talukdar 1