Yexin Li
2025
Learning to Select In-Context Demonstration Preferred by Large Language Model
Zheng Zhang
|
Shaocheng Lan
|
Lei Song
|
Jiang Bian
|
Yexin Li
|
Kan Ren
Findings of the Association for Computational Linguistics: ACL 2025
In-context learning (ICL) enables large language models (LLMs) to adapt to new tasks during inference using only a few demonstrations. However, ICL performance is highly dependent on the selection of these demonstrations. Recent work explores retrieval-based methods for selecting query-specific demonstrations, but these approaches often rely on surrogate objectives such as metric learning, failing to directly optimize ICL performance. Consequently, they struggle to identify truly beneficial demonstrations. Moreover, their discriminative retrieval paradigm is ineffective when the candidate pool lacks sufficient high-quality demonstrations. To address these challenges, we propose GenICL, a novel generative preference learning framework that leverages LLM feedback to directly optimize demonstration selection for ICL. Experiments on 19 datasets across 11 task categories demonstrate that GenICL achieves superior performance than existing methods in selecting the most effective demonstrations, leading to better ICL performance.
Are the Values of LLMs Structurally Aligned with Humans? A Causal Perspective
Yipeng Kang
|
Junqi Wang
|
Yexin Li
|
Mengmeng Wang
|
Wenming Tu
|
Quansen Wang
|
Hengli Li
|
Tingjun Wu
|
Xue Feng
|
Fangwei Zhong
|
Zilong Zheng
Findings of the Association for Computational Linguistics: ACL 2025
As large language models (LLMs) become increasingly integrated into critical applications, aligning their behavior with human values presents significant challenges. Current methods, such as Reinforcement Learning from Human Feedback (RLHF), typically focus on a limited set of coarse-grained values and are resource-intensive. Moreover, the correlations between these values remain implicit, leading to unclear explanations for value-steering outcomes. Our work argues that a latent causal value graph underlies the value dimensions of LLMs and that, despite alignment training, this structure remains significantly different from human value systems. We leverage these causal value graphs to guide two lightweight value-steering methods: role-based prompting and sparse autoencoder (SAE) steering, effectively mitigating unexpected side effects. Furthermore, SAE provides a more fine-grained approach to value steering. Experiments on Gemma-2B-IT and Llama3-8B-IT demonstrate the effectiveness and controllability of our methods.
Search
Fix author
Co-authors
- Jiang Bian 1
- Xue Feng 1
- Yipeng Kang 1
- Shaocheng Lan 1
- Hengli Li 1
- show all...