Yuchen Yan
Other people with similar names: Yuchen Yan
Unverified author pages with similar names: Yuchen Yan
2026
Harnessing Consistency for Robust Test-Time LLM Ensemble
Zhichen Zeng | Qi Yu | Xiao Lin | Ruizhong Qiu | Xuying Ning | Tianxin Wei | Yuchen Yan | Jingrui He | Hanghang Tong
Findings of the Association for Computational Linguistics: EACL 2026
Zhichen Zeng | Qi Yu | Xiao Lin | Ruizhong Qiu | Xuying Ning | Tianxin Wei | Yuchen Yan | Jingrui He | Hanghang Tong
Findings of the Association for Computational Linguistics: EACL 2026
Different large language models (LLMs) exhibit diverse strengths and weaknesses, and LLM ensemble serves as a promising approach to integrate their complementary capabilities. Despite substantial progress in improving ensemble quality, limited attention has been paid to the robustness of ensembles against potential erroneous signals, which often arise from heterogeneous tokenization schemes and varying model expertise. Our analysis shows that ensemble failures typically arise from both the token level and the model level: the former reflects severe disagreement in token predictions, while the latter involves low confidence and pronounced disparities among models. In light of this, we propose CoRE, a plug-and-play technique that harnesses model consistency for robust LLM ensemble, which can be seamlessly integrated with diverse ensemble methods. *Token-level consistency* captures fine-grained disagreements by applying a low-pass filter to downweight uncertain tokens with high inconsistency, often due to token misalignment, thereby improving robustness at a granular level. *Model-level consistency* models global agreement by promoting model outputs with high self-confidence and minimal divergence from others, enhancing robustness at a coarser level. Extensive experiments across diverse benchmarks, model combinations, and ensemble strategies demonstrate that CoRE consistently improves ensemble performance and robustness. Our code is available at https://github.com/zhichenz98/CoRE-EACL26.
2025
To Answer or Not to Answer (TAONA): A Robust Textual Graph Understanding and Question Answering Approach
Yuchen Yan | Aakash Kolekar | Sahika Genc | Wenju Xu | Edward W Huang | Anirudh Srinivasan | Mukesh Jain | Qi He | Hanghang Tong
Findings of the Association for Computational Linguistics: EMNLP 2025
Yuchen Yan | Aakash Kolekar | Sahika Genc | Wenju Xu | Edward W Huang | Anirudh Srinivasan | Mukesh Jain | Qi He | Hanghang Tong
Findings of the Association for Computational Linguistics: EMNLP 2025
Recently, textual graph-based retrieval-augmented generation (GraphRAG) has gained popularity for addressing hallucinations in large language models when answering domain-specific questions. Most existing studies assume that generated answers should comprehensively integrate all relevant information from the textual graph. However, this assumption may not always hold when certain information needs to be vetted or even blocked (e.g., due to safety concerns). In this paper, we target two sides of textual graph understanding and question answering: (1) normal question Answering (A-side): following standard practices, this task generates accurate responses using all relevant information within the textual graph; and (2) Blocked question answering (B-side): A new paradigm where the GraphRAG model must effectively infer and exclude specific relevant information in the generated response. To address these dual tasks, we propose TAONA, a novel GraphRAG model with two variants: (1) TAONA-A for A-side task, which incorporates a specialized GraphEncoder to learn graph prompting vectors; and (2) TAONA-B for B-side task, employing semi-supervised node classification to infer potential blocked graph nodes. Extensive experiments validate TAONA’s superior performance for both A-side and B-side tasks.