Nirav Shah
2026
Knowledge Extraction on Semi-Structured Content: Does It Remain Relevant for Question Answering in the Era of LLMs?
Kai Sun | Yin Huang | Srishti Mehra | Mohammad Kachuee | Xilun Chen | Renjie Tao | Zhaojiang Lin | Andrea Jessee | Nirav Shah | Alex L Betty | Yue Liu | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Kai Sun | Yin Huang | Srishti Mehra | Mohammad Kachuee | Xilun Chen | Renjie Tao | Zhaojiang Lin | Andrea Jessee | Nirav Shah | Alex L Betty | Yue Liu | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
The advent of Large Language Models (LLMs) has significantly advanced web-based Question Answering (QA) systems over semi-structured content, raising questions about the continued utility of knowledge extraction for question answering. This paper investigates the value of triple extraction in this new paradigm by extending an existing benchmark with knowledge extraction annotations and evaluating commercial and open-source LLMs of varying sizes. Our results show that web-scale knowledge extraction remains a challenging task for LLMs. Despite achieving high QA accuracy, LLMs can still benefit from knowledge extraction, through augmentation with extracted triples and multi-task learning. These findings provide insights into the evolving role of knowledge triple extraction in web-based QA and highlight strategies for maximizing LLM effectiveness across different model sizes and resource settings.
2025
PrismRAG: Boosting RAG Factuality with Distractor Resilience and Strategized Reasoning
Mohammad Kachuee | Teja Gollapudi | Minseok Kim | Yin Huang | Kai Sun | Xiao Yang | Jiaqi Wang | Nirav Shah | Yue Liu | Aaron Colak | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Mohammad Kachuee | Teja Gollapudi | Minseok Kim | Yin Huang | Kai Sun | Xiao Yang | Jiaqi Wang | Nirav Shah | Yue Liu | Aaron Colak | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Retrieval-augmented generation (RAG) often falls short when retrieved context includes confusing semi-relevant passages, or when answering questions require deep contextual understanding and reasoning. We propose an efficient fine-tuning framework, called PrismRAG, that (i) trains the model with distractor-aware QA pairs mixing gold evidence with subtle distractor passages, and (ii) instills reasoning-centric habits that make the LLM plan, rationalize, and synthesize without relying on extensive human engineered instructions. Evaluated across 12 open-book RAG QA benchmarks spanning diverse application domains and scenarios, PrismRAG improves average factuality by 5.4%, outperforming state-of-the-art solutions. Our method is being deployed in production.