2025
pdf
bib
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
Ashutosh Modi
|
Saptarshi Ghosh
|
Asif Ekbal
|
Pawan Goyal
|
Sarika Jain
|
Abhinav Joshi
|
Shivani Mishra
|
Debtanu Datta
|
Shounak Paul
|
Kshetrimayum Boynao Singh
|
Sandeep Kumar
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
pdf
bib
abs
Overview of the 1st Workshop on NLP for Empowering Justice
Ashutosh Modi
|
Saptarshi Ghosh
|
Asif Ekbal
|
Pawan Goyal
|
Sarika Jain
|
Abhinav Joshi
|
Shivani Mishra
|
Debtanu Datta
|
Shounak Paul
|
Kshetrimayum Boynao Singh
|
Sandeep Kumar
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
The first iteration of the JUST-NLP: Workshop on NLP for Empowering Justice was organized to accelerate research in Natural Language Processing for legal text processing. The inaugural edition, JUST-NLP 2025, was held as a hybrid event at IJCNLP-AACL 2025 on December 24 at IIT Bombay. The program featured a research track, four invited talks, and two shared tasks: (1) L-SUMM, an abstractive summarization task for Indian legal judgments, and (2) L-MT, a legal machine translation task between English and Hindi. The workshop received strong interest from the community, with 29 submissions, of which 21 were accepted. Among the accepted papers, 5 were regular research-track papers published in the proceedings, and 2 were accepted as non-archival presentations. For the shared tasks, 9 papers were accepted for L-SUMM, and 5 papers were accepted for L-MT, for publication in the proceedings. The workshop focused on a broad set of Legal NLP challenges, including information extraction, retrieval, multilingual processing, legal reasoning, and applications of large language models. Overall, JUST-NLP 2025 aimed to bring together AI researchers and legal practitioners to develop scalable, domain-aware NLP methods that can support legal workflows and contribute toward more efficient and equitable justice systems.
pdf
bib
abs
Findings of the JUST-NLP 2025 Shared Task on Summarization of Indian Court Judgments
Debtanu Datta
|
Shounak Paul
|
Kshetrimayum Boynao Singh
|
Sandeep Kumar
|
Abhinav Joshi
|
Shivani Mishra
|
Sarika Jain
|
Asif Ekbal
|
Pawan Goyal
|
Ashutosh Modi
|
Saptarshi Ghosh
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
This paper presents an overview of the Shared Task on Summarization of Indian Court Judgments (L-SUMM), hosted by the JUST-NLP 2025 Workshop at IJCNLP-AACL 2025. This task aims to increase research interest in automatic summarization techniques for lengthy and intricate legal documents from the Indian judiciary. It particularly addresses court judgments that contain dense legal reasoning and semantic roles that must be preserved in summaries. As part of this shared task, we introduce the Indian Legal Summarization (L-SUMM) dataset, comprising 1,800 Indian court judgments paired with expert-written abstractive summaries, both in English. Therefore, the task focuses on generating high-quality abstractive summaries of court judgments in English. A total of 9 teams participated in this task, exploring a diverse range of methodologies, including transformer-based models, extractive-abstractive hybrids, graph-based ranking approaches, long-context LLMs, and rhetorical-role-based techniques. This paper describes the task setup, dataset, evaluation framework, and our findings. We report the results and highlight key trends across participant approaches, including the effectiveness of hybrid pipelines and challenges in handling extreme sequence lengths.
pdf
bib
abs
Findings of the JUST-NLP 2025 Shared Task on English-to-Hindi Legal Machine Translation
Kshetrimayum Boynao Singh
|
Sandeep Kumar
|
Debtanu Datta
|
Abhinav Joshi
|
Shivani Mishra
|
Shounak Paul
|
Pawan Goyal
|
Sarika Jain
|
Saptarshi Ghosh
|
Ashutosh Modi
|
Asif Ekbal
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
This paper provides an overview of the Shared Task on Legal Machine Translation (L-MT), organized as part of the JUST-NLP 2025 Workshop at IJCNLP-AACL 2025, aimed at improving the translation of legal texts, a domain where precision, structural faithfulness, and terminology preservation are essential. The training set comprises 50,000 sentences, with 5,000 sentences each for the validation and test sets. The submissions employed strategies such as: domain-adaptive fine-tuning of multilingual models, QLoRA-based parameter-efficient adaptation, curriculum-guided supervised training, reinforcement learning with verifiable MT metrics, and from-scratch Transformer training. The systems are evaluated based on BLEU, METEOR, TER, chrF++, BERTScore, and COMET metrics. We also combine the scores of these metrics to give an average score (AutoRank). The top-performing system is based on a fine-tuned distilled NLLB-200 model and achieved the highest AutoRank score of 72.1. Domain adaptation consistently yielded substantial improvements over baseline models, and precision-focused rewards proved especially effective for the legal MT. The findings also highlight that large multilingual Transformers can deliver accurate and reliable English-to-Hindi legal translations when carefully fine-tuned on legal data, advancing the broader goal of improving access to justice in multilingual settings.
pdf
bib
abs
LeCNet: A Legal Citation Network Benchmark Dataset
Pooja Harde
|
Bhavya Jain
|
Sarika Jain
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
Legal document analysis is pivotal in modern judicial systems, particularly for case retrieval, classification, and recommendation tasks. Graph neural networks (GNNs) have revolutionized legal use cases by enabling the efficient analysis of complex relationships. Although existing legal citation network datasets have significantly advanced research in this domain, the lack of large-scale open-source datasets tailored to the Indian judicial system has limited progress. To address this gap, we present the Indian Legal Citation Network (LeCNet) - the first open-source benchmark dataset for the link prediction task (missing citation recommendation) in the Indian judicial context. The dataset has been created by extracting information from the original judgments. LeCNet comprises 26,308 nodes representing case judgments and 67,108 edges representing citation relationships between the case nodes. Each node is described with rich features of document embeddings that incorporate contextual information from the case documents. Baseline experiments using various machine learning models were conducted for dataset validation. The Mean Reciprocal Rank (MRR) metric is used for model evaluation. The results obtained demonstrate the utility of the LeCNet dataset, highlighting the advantages of graph-based representations over purely textual models.
pdf
bib
abs
Legal Document Summarization: A Zero-shot Modular Agentic Workflow Approach
Taha Sadikot
|
Sarika Jain
Proceedings of the 1st Workshop on NLP for Empowering Justice (JUST-NLP 2025)
The increasing volume and complexity of Indian High Court judgments require high-quality automated summarization systems. Our agentic workflow framework for the summarization of Indian High Court judgments achieves competitive results without model fine-tuning. Experiments on CivilSum and IN-Abs test sets report ROUGE-1 F1 up to 0.547 and BERTScore F1 up to 0.866, comparable to state-of-the-art supervised models, with advantages in transparency and efficiency. We introduce two zero-shot modular agentic workflows: Lexical Modular Summarizer (LexA), a three-stage modular architecture optimized for lexical overlap (ROUGE), and Semantic Agentic Summarizer (SemA), a five-stage integrated architecture optimized for semantic similarity (BERTScore). Both workflows operate without supervised model fine-tuning, instead relying on strategic data processing, modular agent orchestration, and carefully engineered prompts. Our framework achieves ROUGE-1 F1 of 0.6326 and BERTScore F1 of 0.8902 on CivilSum test set, and ROUGE-1 F1 of 0.1951 and BERTScore F1 of 0.8299 on IN-Abs test set, substantially outperforming zero-shot baselines, rivaling leading fine-tuned transformer models while requiring no supervised training. This work demonstrates that modular, zero-shot agentic approaches can deliver production-grade results for legal summarization, offering a new direction for resource-limited judicial settings.