2025
pdf
bib
abs
Advancing Uto-Aztecan Language Technologies: A Case Study on the Endangered Comanche Language
Jesus Alvarez C
|
Daua Karajeanes
|
Ashley Prado
|
John Ruttan
|
Ivory Yang
|
Sean O’brien
|
Vasu Sharma
|
Kevin Zhu
Proceedings of the Fifth Workshop on NLP for Indigenous Languages of the Americas (AmericasNLP)
The digital exclusion of endangered languages remains a critical challenge in NLP, limiting both linguistic research and revitalization efforts. This study introduces the first computational investigation of Comanche, an Uto-Aztecan language on the verge of extinction, demonstrating how minimal-cost, community-informed NLP interventions can support language preservation. We present a manually curated dataset of 412 phrases, a synthetic data generation pipeline, and an empirical evaluation of GPT-4o and GPT-4o-mini for language identification. Our experiments reveal that while LLMs struggle with Comanche in zero-shot settings, few-shot prompting significantly improves performance, achieving near-perfect accuracy with just five examples. Our findings highlight the potential of targeted NLP methodologies in low-resource contexts and emphasize that visibility is the first step toward inclusion. By establishing a foundation for Comanche in NLP, we advocate for computational approaches that prioritize accessibility, cultural sensitivity, and community engagement.
pdf
bib
abs
Rosetta-PL: Propositional Logic as a Benchmark for Large Language Model Reasoning
Shaun Lee Baek
|
Shaun Esua-Mensah
|
Cyrus Tsui
|
Sejan Vigneswaralingam
|
Abdullah Alali
|
Michael Lu
|
Vasu Sharma
|
Kevin Zhu
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 4: Student Research Workshop)
Large Language Models (LLMs) are primarily trained on high-resource natural languages, limiting their effectiveness in low-resource settings and in tasks requiring deep logical reasoning. This research introduces Rosetta-PL, a benchmark designed to evaluate LLMs’ logical reasoning and generalization capabilities in a controlled environment. We construct Rosetta-PL by translating a dataset of logical propositions from Lean into a custom logical language, which is then used to fine-tune an LLM (e.g., GPT-4o). Our experiments analyze the impact of the size of the dataset and the translation methodology on the performance of the model. Our results indicate that preserving logical relationships in the translation process significantly boosts precision, with accuracy plateauing beyond roughly 20,000 training samples. These insights provide valuable guidelines for optimizing LLM training in formal reasoning tasks and improving performance in various low-resource language applications.
2022
pdf
bib
abs
Tweet Based Reach Aware Temporal Attention Network for NFT Valuation
Ramit Sawhney
|
Megh Thakkar
|
Ritesh Soun
|
Atula Neerkaje
|
Vasu Sharma
|
Dipanwita Guhathakurta
|
Sudheer Chava
Findings of the Association for Computational Linguistics: EMNLP 2022
Non-Fungible Tokens (NFTs) are a relatively unexplored class of assets. Designing strategies to forecast NFT trends is an intricate task due to its extremely volatile nature. The market is largely driven by public sentiment and “hype”, which in turn has a high correlation with conversations taking place on social media platforms like Twitter. Prior work done for modelling stock market data does not take into account the extent of impact certain highly influential tweets and their authors can have on the market. Building on these limitations and the nature of the NFT market, we propose a novel reach-aware temporal learning approach to make predictions for forecasting future trends in the NFT market. We perform experiments on a new dataset consisting of over 1.3 million tweets and 180 thousand NFT transactions spanning over 15 NFT collections curated by us. Our model (TA-NFT) outperforms other state-of-the-art methods by an average of 36%. Through extensive quantitative and ablative analysis, we demonstrate the ability of our approach as a practical method for predicting NFT trends.
2018
pdf
bib
abs
BioAMA: Towards an End to End BioMedical Question Answering System
Vasu Sharma
|
Nitish Kulkarni
|
Srividya Pranavi
|
Gabriel Bayomi
|
Eric Nyberg
|
Teruko Mitamura
Proceedings of the BioNLP 2018 workshop
In this paper, we present a novel Biomedical Question Answering system, BioAMA: “Biomedical Ask Me Anything” on task 5b of the annual BioASQ challenge. In this work, we focus on a wide variety of question types including factoid, list based, summary and yes/no type questions that generate both exact and well-formed ‘ideal’ answers. For summary-type questions, we combine effective IR-based techniques for retrieval and diversification of relevant snippets for a question to create an end-to-end system which achieves a ROUGE-2 score of 0.72 and a ROUGE-SU4 score of 0.71 on ideal answer questions (7% improvement over the previous best model). Additionally, we propose a novel NLI-based framework to answer the yes/no questions. To train the NLI model, we also devise a transfer-learning technique by cross-domain projection of word embeddings. Finally, we present a two-stage approach to address the factoid and list type questions by first generating a candidate set using NER taggers and ranking them using both supervised or unsupervised techniques.
pdf
bib
abs
Cyclegen: Cyclic consistency based product review generator from attributes
Vasu Sharma
|
Harsh Sharma
|
Ankita Bishnu
|
Labhesh Patel
Proceedings of the 11th International Conference on Natural Language Generation
In this paper we present an automatic review generator system which can generate personalized reviews based on the user identity, product identity and designated rating the user wishes to allot to the review. We combine this with a sentiment analysis system which performs the complimentary task of assigning ratings to reviews based purely on the textual content of the review. We introduce an additional loss term to ensure cyclic consistency of the sentiment rating of the generated review with the conditioning rating used to generate the review. The introduction of this new loss term constraints the generation space while forcing it to generate reviews adhering better to the requested rating. The use of ‘soft’ generation and cyclic consistency allows us to train our model in an end to end fashion. We demonstrate the working of our model on product reviews from Amazon dataset.
2017
pdf
bib
Segmentation Guided Attention Networks for Visual Question Answering
Vasu Sharma
|
Ankita Bishnu
|
Labhesh Patel
Proceedings of ACL 2017, Student Research Workshop
2016
pdf
bib
Automatic tagging and retrieval of E-Commerce products based on visual features
Vasu Sharma
|
Harish Karnick
Proceedings of the NAACL Student Research Workshop
2015
pdf
bib
Analyzing Newspaper Crime Reports for Identification of Safe Transit Paths
Vasu Sharma
|
Rajat Kulshreshtha
|
Puneet Singh
|
Nishant Agrawal
|
Akshay Kumar
Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Student Research Workshop