Haifeng Chen
Other people with similar names: Haifeng Chen
Unverified author pages with similar names: Haifeng Chen
2026
Multi-Agent Procedural Graph Extraction with Structural and Logical Refinement
Wangyang Ying | Yanchi Liu | Xujiang Zhao | Wei Cheng | Zhengzhang Chen | Wenchao Yu | Yanjie Fu | Haifeng Chen
Findings of the Association for Computational Linguistics: EACL 2026
Wangyang Ying | Yanchi Liu | Xujiang Zhao | Wei Cheng | Zhengzhang Chen | Wenchao Yu | Yanjie Fu | Haifeng Chen
Findings of the Association for Computational Linguistics: EACL 2026
Automatically extracting workflows as procedural graphs from natural language is a promising yet underexplored task that requires ensuring both structural validity and logical alignment. Recent advances in large language models (LLMs) show potential for graph extraction, but often yield ill-formed structures or misinterpret logical constructs such as gateways. We introduce , a multi-agent framework that treats procedural graph extraction as a multi-round reasoning process with structural and logical refinement agents. The framework operates in three iterative stages: (1) an LLM-based graph extraction phase, (2) a structural feedback phase where a simulation agent diagnoses and explains structural issues, and (3) a logical feedback phase where a semantic agent aligns semantics between flow logic and linguistic cues in the source text. Important feedback is prioritized and expressed in natural language, which is injected into the next-round prompt, enabling interpretable and controllable refinement. This modular design allows agents to target distinct error types without supervision or parameter updates. Experiments demonstrate that achieves substantial improvements in both structural correctness and logical consistency over strong baselines.
DeepSieve: Information Sieving via LLM-as-a-Knowledge-Router
Minghao Guo | Qingcheng Zeng | Xujiang Zhao | Yanchi Liu | Wenchao Yu | Mengnan Du | Haifeng Chen | Wei Cheng
Findings of the Association for Computational Linguistics: EACL 2026
Minghao Guo | Qingcheng Zeng | Xujiang Zhao | Yanchi Liu | Wenchao Yu | Mengnan Du | Haifeng Chen | Wei Cheng
Findings of the Association for Computational Linguistics: EACL 2026
Large Language Models (LLMs) excel at many reasoning tasks but struggle with knowledge-intensive queries due to their inability to dynamically access up-to-date or domain-specific information. Retrieval-Augmented Generation (RAG) has emerged as a promising solution, enabling LLMs to ground their responses in external sources. However, existing RAG methods lack fine-grained control over both the query and source sides, often resulting in noisy retrieval and shallow reasoning. In this work, we introduce DeepSieve, an agentic RAG framework that incorporates information sieving via LLM-as-a-knowledge-router. DeepSieve decomposes complex queries into structured sub-questions and recursively routes each to the most suitable knowledge source, filtering irrelevant information through a multi-stage distillation process. Our design emphasizes modularity, transparency, and adaptability, leveraging recent advances in agentic system design. Experiments on multi-hop QA tasks across heterogeneous sources demonstrate improved reasoning depth, retrieval precision, and interpretability over conventional RAG approaches.
Decoding Time Series with LLMs: A Multi-Agent Framework for Cross-Domain Annotation
Minhua Lin | Zhengzhang Chen | Yanchi Liu | Xujiang Zhao | Zongyu Wu | Junxiang Wang | Xiang Zhang | Suhang Wang | Haifeng Chen
Findings of the Association for Computational Linguistics: EACL 2026
Minhua Lin | Zhengzhang Chen | Yanchi Liu | Xujiang Zhao | Zongyu Wu | Junxiang Wang | Xiang Zhang | Suhang Wang | Haifeng Chen
Findings of the Association for Computational Linguistics: EACL 2026
Time series data is ubiquitous across various domains, including manufacturing, finance, and healthcare. High-quality annotations are essential for effectively understanding time series and facilitating downstream tasks. However, obtaining such annotations is challenging, particularly in mission-critical domains. In this paper, we propose TESSA, a multi-agent system designed to automatically generate both general and domain-specific annotations for time series data. TESSA introduces two agents: a general annotation agent and a domain-specific annotation agent. The general agent captures common patterns and knowledge across multiple source domains, leveraging both time-series-wise and text-wise features to generate general annotations. Meanwhile, the domain-specific agent utilizes limited annotations from the target domain to learn domain-specific terminology and generate targeted annotations. Extensive experiments on multiple synthetic and real-world datasets demonstrate that TESSA effectively generates high-quality annotations, outperforming existing methods.