Yin Huang
2026
Knowledge Extraction on Semi-Structured Content: Does It Remain Relevant for Question Answering in the Era of LLMs?
Kai Sun | Yin Huang | Srishti Mehra | Mohammad Kachuee | Xilun Chen | Renjie Tao | Zhaojiang Lin | Andrea Jessee | Nirav Shah | Alex L Betty | Yue Liu | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Kai Sun | Yin Huang | Srishti Mehra | Mohammad Kachuee | Xilun Chen | Renjie Tao | Zhaojiang Lin | Andrea Jessee | Nirav Shah | Alex L Betty | Yue Liu | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
The advent of Large Language Models (LLMs) has significantly advanced web-based Question Answering (QA) systems over semi-structured content, raising questions about the continued utility of knowledge extraction for question answering. This paper investigates the value of triple extraction in this new paradigm by extending an existing benchmark with knowledge extraction annotations and evaluating commercial and open-source LLMs of varying sizes. Our results show that web-scale knowledge extraction remains a challenging task for LLMs. Despite achieving high QA accuracy, LLMs can still benefit from knowledge extraction, through augmentation with extracted triples and multi-task learning. These findings provide insights into the evolving role of knowledge triple extraction in web-based QA and highlight strategies for maximizing LLM effectiveness across different model sizes and resource settings.
Aligning Paralinguistic Understanding and Generation in Speech LLMs via Multi-Task Reinforcement Learning
Minseok Kim | Jingxiang Chen | Seong-Gyun Leem | Yin Huang | Rashi Rungta | Zhicheng Ouyang | Haibin Wu | Surya Teja Appini | Ankur Bansal | Yang Bai | Yue Liu | Florian Metze | Ahmed A Aly | Anuj Kumar | Ariya Rastrow | Zhaojiang Lin
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 5: Industry Track)
Minseok Kim | Jingxiang Chen | Seong-Gyun Leem | Yin Huang | Rashi Rungta | Zhicheng Ouyang | Haibin Wu | Surya Teja Appini | Ankur Bansal | Yang Bai | Yue Liu | Florian Metze | Ahmed A Aly | Anuj Kumar | Ariya Rastrow | Zhaojiang Lin
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 5: Industry Track)
Speech large language models (LLMs) observe paralinguistic cues such as prosody, emotion, and non-verbal sounds—crucial for intent understanding. However, leveraging these cues faces challenges: limited training data, annotation difficulty, and models exploiting lexical shortcuts over paralinguistic signals. We propose multi-task reinforcement learning (RL) with chain-of-thought prompting that elicits explicit affective reasoning. To address data scarcity, we introduce a paralinguistics-aware speech LLM (PALLM) that jointly optimizes sentiment classification from audio and paralinguistics-aware response generation via a two-stage pipeline. Experiments demonstrate that our approach improves paralinguistics understanding over both supervised baselines and strong proprietary models (Gemini-2.5-Pro, GPT-4o-audio), by 8-12% on Expresso, IEMOCAP, and RAVDESS. The results show that modeling paralinguistic reasoning with multi-task RL is crucial for building emotionally intelligent speech LLMs.
2025
PrismRAG: Boosting RAG Factuality with Distractor Resilience and Strategized Reasoning
Mohammad Kachuee | Teja Gollapudi | Minseok Kim | Yin Huang | Kai Sun | Xiao Yang | Jiaqi Wang | Nirav Shah | Yue Liu | Aaron Colak | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Mohammad Kachuee | Teja Gollapudi | Minseok Kim | Yin Huang | Kai Sun | Xiao Yang | Jiaqi Wang | Nirav Shah | Yue Liu | Aaron Colak | Anuj Kumar | Wen-tau Yih | Xin Luna Dong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track
Retrieval-augmented generation (RAG) often falls short when retrieved context includes confusing semi-relevant passages, or when answering questions require deep contextual understanding and reasoning. We propose an efficient fine-tuning framework, called PrismRAG, that (i) trains the model with distractor-aware QA pairs mixing gold evidence with subtle distractor passages, and (ii) instills reasoning-centric habits that make the LLM plan, rationalize, and synthesize without relying on extensive human engineered instructions. Evaluated across 12 open-book RAG QA benchmarks spanning diverse application domains and scenarios, PrismRAG improves average factuality by 5.4%, outperforming state-of-the-art solutions. Our method is being deployed in production.
Search
Fix author
Co-authors
- Anuj Kumar 3
- Yue Liu 3
- Xin Luna Dong 2
- Mohammad Kachuee 2
- Minseok Kim 2
- Zhaojiang Lin 2
- Nirav Shah 2
- Kai Sun 2
- Wen-tau Yih 2
- Ahmed A Aly 1
- Surya Teja Appini 1
- Yang Bai 1
- Ankur Bansal 1
- Alex L Betty 1
- Xilun Chen 1
- Jingxiang Chen 1
- Aaron Colak 1
- Teja Gollapudi 1
- Andrea Jessee 1
- Seong-Gyun Leem 1
- Srishti Mehra 1
- Florian Metze 1
- Zhicheng Ouyang 1
- Ariya Rastrow 1
- Rashi Rungta 1
- Renjie Tao 1
- Jiaqi Wang 1
- Haibin Wu 1
- Xiao Yang 1