Heyuan Huang
May refer to several people
Other people with similar names: Heyuan Huang (JHU)
2025
RAG+: Enhancing Retrieval-Augmented Generation with Application-Aware Reasoning
Yu Wang
|
Shiwan Zhao
|
Zhihu Wang
|
Ming Fan
|
Xicheng Zhang
|
Yubo Zhang
|
Zhengfan Wang
|
Heyuan Huang
|
Ting Liu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
The integration of external knowledge through Retrieval-Augmented Generation (RAG) has become foundational in enhancing large language models (LLMs) for knowledge-intensive tasks. However, existing RAG paradigms often overlook the cognitive step of applying knowledge, leaving a gap between retrieved facts and task-specific reasoning. In this work, we introduce RAG+, a principled and modular extension that explicitly incorporates application-aware reasoning into the RAG pipeline. RAG+ constructs a dual corpus consisting of knowledge and aligned application examples, created either manually or automatically, and jointly retrieves both during inference. This design enables LLMs not only to access relevant information but also to apply it within structured, goal-oriented reasoning processes. Experiments across mathematical, law, and medical domains, conducted on multiple models, demonstrate that RAG+ consistently outperforms standard RAG variants, achieving average improvements of 3–5%, and peak gains up to 13.5% in complex scenarios. By bridging retrieval with actionable application, RAG+ advances a more cognitively grounded framework for knowledge integration, representing a step toward more interpretable and capable LLMs.
Re-TASK: Revisiting LLM Tasks from Capability, Skill, and Knowledge Perspectives
Zhihu Wang
|
Shiwan Zhao
|
Yu Wang
|
Heyuan Huang
|
Sitao Xie
|
Yubo Zhang
|
Jiaxin Shi
|
Zhixing Wang
|
Hongyan Li
|
Junchi Yan
Findings of the Association for Computational Linguistics: ACL 2025
The Chain-of-Thought (CoT) paradigm has become a pivotal method for solving complex problems with large language models (LLMs). However, its application to domain-specific tasks remains challenging, as LLMs often fail to decompose tasks accurately or execute subtasks effectively. This paper introduces the Re-TASK framework, a novel theoretical model that Revisits LLM Tasks from cApability, Skill, and Knowledge perspectives, drawing on the principles of Bloom’s Taxonomy and Knowledge Space Theory. While CoT provides a workflow-centric perspective on tasks, Re-TASK introduces a Chain-of-Learning (CoL) paradigm that highlights task dependencies on specific capability items, further broken down into their constituent knowledge and skill components. To address CoT failures, we propose a Re-TASK prompting strategy, which strengthens task-relevant capabilities through targeted knowledge injection and skill adaptation. Experiments across diverse domains demonstrate the effectiveness of Re-TASK. In particular, we achieve improvements of 45.00% on Yi-1.5-9B and 24.50% on Llama3-Chinese-8B for legal tasks. These results highlight the potential of Re-TASK to significantly enhance LLM performance and its applicability in specialized domains. We release our code and data at https://github.com/Uylee/Re-TASK.
Search
Fix author
Co-authors
- Yu Wang (王昱, 王雨) 2
- Zhihu Wang 2
- Yubo Zhang 2
- Shiwan Zhao 2
- Ming Fan 1
- show all...