Devasha Trivedi
2026
Reasoning Graph-Structured Question Answering: Datasets and Insights from LLM Benchmarking
Khin Yone | Devasha Trivedi | Anish Pahilajani | Jincen Shuai | Samyak Rajesh Jain | Ryan Rossi | Nesreen K. Ahmed | Franck Dernoncourt | Yu Wang | Namyong Park
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Khin Yone | Devasha Trivedi | Anish Pahilajani | Jincen Shuai | Samyak Rajesh Jain | Ryan Rossi | Nesreen K. Ahmed | Franck Dernoncourt | Yu Wang | Namyong Park
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Large Language Models (LLMs) have shown remarkable success in multi-hop question-answering (M-QA) due to their advanced reasoning capabilities. However, the influence of reasoning structures on their performance remains underexplored, primarily due to the lack of M-QA datasets that explicitly encode the reasoning pathways underlying each question-answer pair. To address this gap, we introduce the reasoning graph-structured question answering dataset (GRS-QA), which provides both semantic contexts and reasoning structures for the QA pairs. Unlike existing M-QA datasets, GRS-QA explicitly captures intricate reasoning pathways through reasoning graphs, where nodes correspond to textual contexts and edges denote logical flows. Using GRS-QA, we systematically evaluate LLM performance across varying context structures, prompting styles, and data domains. Our empirical analysis reveals that LLMs perform differently based on the reasoning structure, context, and prompting styles, indicating their varying ability to leverage graph-structured knowledge. Notably, providing explicit reasoning guidance proves more effective than supplying contextual information alone.
2024
NLP at UC Santa Cruz at SemEval-2024 Task 5: Legal Answer Validation using Few-Shot Multi-Choice QA
Anish Pahilajani | Samyak Jain | Devasha Trivedi
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)
Anish Pahilajani | Samyak Jain | Devasha Trivedi
Proceedings of the 18th International Workshop on Semantic Evaluation (SemEval-2024)
This paper presents our submission to the SemEval 2024 Task 5: The Legal Argument Reasoning Task in Civil Procedure. We present two approaches to solving the task of legal answer validation, given an introduction to the case, a question and an answer candidate. Firstly, we fine-tuned pre-trained BERT-based models and found that models trained on domain knowledge perform better. Secondly, we performed few-shot prompting on GPT models and found that reformulating the answer validation task to be a multiple-choice QA task remarkably improves the performance of the model. Our best submission is a BERT-based model that achieved the 7th place out of 20.