Yang Liu
Other people with similar names:
Yang Liu
,
Yang Liu
(Wilfrid Laurier University),
Yang Liu (刘扬)
(刘扬; Ph.D Purdue; ICSI, Dallas, Facebook, Liulishuo, Amazon),
Yang Liu
,
Yang Liu (刘洋)
(刘洋; ICT, Tsinghua, Beijing Academy of Artificial Intelligence),
Yang Liu
(Edinburgh Ph.D., Microsoft),
Yang Liu
(University of Helsinki),
Yang Liu
(Samsung Research Center Beijing),
Yang Liu
(Tianjin University, China),
Yang Liu
,
Yang Liu
(Microsoft Cognitive Services Research),
Yang Liu
(Univ. of Michigan, UC Santa Cruz),
Yang Liu
(National University of Defense Technology),
Yang Liu
,
Yang Liu
,
Yang Liu
,
Yang Janet Liu
(Georgetown University; 刘洋),
Yang Liu (刘扬)
(Peking University),
Yang Liu
(The Chinese University of Hong Kong (Shenzhen)),
Yang Liu
,
Yang Liu
,
Yang Liu
(3M Health Information Systems),
Yang Liu
(Beijing Language and Culture University)
2025
pdf
bib
abs
Understanding and Leveraging the Expert Specialization of Context Faithfulness in Mixture-of-Experts LLMs
Jun Bai
|
Minghao Tong
|
Yang Liu
|
Zixia Jia
|
Zilong Zheng
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Context faithfulness is essential for reliable reasoning in context-dependent scenarios. However, large language models often struggle to ground their outputs in the provided context, resulting in irrelevant responses.Inspired by the emergent expert specialization observed in mixture-of-experts architectures, this work investigates whether certain experts exhibit specialization in context utilization—offering a potential pathway toward targeted optimization for improved context faithfulness.To explore this, we propose Router Lens, a method that accurately identifies context-faithful experts. Our analysis reveals that these experts progressively amplify attention to relevant contextual information, thereby enhancing context grounding.Building on this insight, we introduce Context-faithful Expert Fine-Tuning (CEFT), a lightweight optimization approach that selectively fine-tunes context-faithful experts.Experiments across a wide range of benchmarks and models demonstrate that CEFT matches or surpasses the performance of full fine-tuning while being significantly more efficient.
pdf
bib
abs
ReflectEvo: Improving Meta Introspection of Small LLMs by Learning Self-Reflection
Jiaqi Li
|
Xinyi Dong
|
Yang Liu
|
Zhizhuo Yang
|
Quansen Wang
|
Xiaobo Wang
|
Song-Chun Zhu
|
Zixia Jia
|
Zilong Zheng
Findings of the Association for Computational Linguistics: ACL 2025
We present a novel pipeline, ReflectEvo, to demonstrate that small language models (SLMs) can enhance meta introspection through reflection learning. This process iteratively generates self-reflection for self-training, fostering a continuous and self-evolving process. Leveraging this pipeline, we construct ReflectEvo-460k, a large-scale, comprehensive, self-generated reflection dataset with broadened instructions and diverse multi-domain tasks. Building upon this dataset, we demonstrate the effectiveness of reflection learning to improve SLMs’ reasoning abilities using SFT and DPO with remarkable performance, substantially boosting Llama-3 from 52.4% to 71.2% and Mistral from 44.4% to 71.1%. It validates that ReflectEvo can rival or even surpass the reasoning capability of the three prominent open-sourced models on BIG-bench without distillation from superior models or fine-grained human annotation. We further conduct a deeper analysis of the high quality of self-generated reflections and their impact on error localization and correction. Our work highlights the potential of continuously enhancing the reasoning performance of SLMs through iterative reflection learning in the long run.