Yuyao Zhang
2025
Hierarchical Document Refinement for Long-context Retrieval-augmented Generation
Jiajie Jin
|
Xiaoxi Li
|
Guanting Dong
|
Yuyao Zhang
|
Yutao Zhu
|
Yongkang Wu
|
Zhonghua Li
|
Ye Qi
|
Zhicheng Dou
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Real-world RAG applications often encounter long-context input scenarios, where redundant information and noise results in higher inference costs and reduced performance. To address these challenges, we propose LongRefiner, an efficient plug-and-play refiner that leverages the inherent structural characteristics of long documents. LongRefiner employs dual-level query analysis, hierarchical document structuring, and adaptive refinement through multi-task learning on a single foundation model. Experiments on seven QA datasets demonstrate that LongRefiner achieves competitive performance in various scenarios while using 10x fewer computational costs and latency compared to the best baseline. Further analysis validates that LongRefiner is scalable, efficient, and effective, providing practical insights for real-world long-text RAG applications. Our code is available at https://github.com/ignorejjj/LongRefiner.
Neuro-Symbolic Query Compiler
Yuyao Zhang
|
Zhicheng Dou
|
Xiaoxi Li
|
Jiajie Jin
|
Yongkang Wu
|
Zhonghua Li
|
Ye Qi
|
Ji-Rong Wen
Findings of the Association for Computational Linguistics: ACL 2025
Precise recognition of search intent in Retrieval-Augmented Generation (RAG) systems remains a challenging goal, especially under resource constraints and for complex queries with nested structures and dependencies. This paper presents **QCompiler**, a neuro-symbolic framework inspired by linguistic grammar rules and compiler design, to bridge this gap. It theoretically presents a minimal yet sufficient Backus-Naur Form (BNF) grammar G[q] to formalize complex queries. Unlike previous methods, this grammar maintains completeness while minimizing redundancy. Based on this, QCompiler includes a query expression translator, a Lexical syntax parser, and a Recursive Descent Processor to compile queries into Abstract Syntax Trees (ASTs) for execution. The atomicity of the sub-queries in the leaf nodes ensures more precise document retrieval and response generation, significantly improving the RAG system’s ability to address complex queries.
2024
Enabling Discriminative Reasoning in LLMs for Legal Judgment Prediction
Chenlong Deng
|
Kelong Mao
|
Yuyao Zhang
|
Zhicheng Dou
Findings of the Association for Computational Linguistics: EMNLP 2024
Legal judgment prediction is essential for enhancing judicial efficiency. In this work, we identify that existing large language models (LLMs) underperform in this domain due to challenges in understanding case complexities and distinguishing between similar charges. To adapt LLMs for effective legal judgment prediction, we introduce the Ask-Discriminate-Predict (ADAPT) reasoning framework inspired by human judicial reasoning. ADAPT involves decomposing case facts, discriminating among potential charges, and predicting the final judgment. We further enhance LLMs through fine-tuning with multi-task synthetic trajectories to improve legal judgment prediction accuracy and efficiency under our ADAPT framework. Extensive experiments conducted on two widely-used datasets demonstrate the superior performance of our framework in legal judgment prediction, particularly when dealing with complex and confusing charges.
Search
Fix author
Co-authors
- Zhicheng Dou (窦志成) 3
- Jiajie Jin 2
- Xiaoxi Li 2
- Zhonghua Li 2
- Ye Qi 2
- show all...