Haozhu Wang
2025
A Systematic Survey of Automatic Prompt Optimization Techniques
Kiran Ramnath
|
Kang Zhou
|
Sheng Guan
|
Soumya Smruti Mishra
|
Xuan Qi
|
Zhengyuan Shen
|
Shuai Wang
|
Sangmin Woo
|
Sullam Jeoung
|
Yawei Wang
|
Haozhu Wang
|
Han Ding
|
Yuzhe Lu
|
Zhichao Xu
|
Yun Zhou
|
Balasubramaniam Srinivasan
|
Qiaojing Yan
|
Yueyan Chen
|
Haibo Ding
|
Panpan Xu
|
Lin Lee Cheong
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Since the advent of large language models (LLMs), prompt engineering has been a crucial step for eliciting desired responses for various Natural Language Processing (NLP) tasks. However, prompt engineering remains an impediment for end users due to rapid advances in models, tasks, and associated best practices. To mitigate this, Automatic Prompt Optimization (APO) techniques have recently emerged that use various automated techniques to help improve the performance of LLMs on various tasks. In this paper, we present a comprehensive survey summarizing the current progress and remaining challenges in this field. We provide a formal definition of APO, a 5-part unifying framework, and then proceed to rigorously categorize all relevant works based on their salient features therein. We hope to spur further research guided by our framework.
2024
LaRS: Latent Reasoning Skills for Chain-of-Thought Reasoning
Zifan Xu
|
Haozhu Wang
|
Dmitriy Bespalov
|
Xian Wu
|
Peter Stone
|
Yanjun Qi
Findings of the Association for Computational Linguistics: EMNLP 2024
Chain-of-thought (CoT) prompting is a popular in-context learning (ICL) approach for large language models (LLMs), especially when tackling complex reasoning tasks. Traditional ICL approaches construct prompts using examples that contain questions similar to the input question. However, CoT prompting, which includes crucial intermediate reasoning steps (rationales) within its examples, necessitates selecting examples based on these rationales rather than the questions themselves. Existing methods require human experts or pre-trained LLMs to describe the skill, a high-level abstraction of rationales, to guide the selection. These methods, however, are often costly and difficult to scale. Instead, this paper introduces a new approach named Latent Reasoning Skills (LaRS) that employs unsupervised learning to create a latent space representation of rationales, with a latent variable called a reasoning skill. Concurrently, LaRS learns a reasoning policy to determine the required reasoning skill for a given question. Then the ICL examples are selected by aligning the reasoning skills between past examples and the question. This approach is theoretically grounded and compute-efficient, eliminating the need for auxiliary LLM inference or manual prompt design. Empirical results demonstrate that LaRS consistently outperforms SOTA skill-based selection methods, processing example banks four times faster, reducing LLM inferences during the selection stage by half, and showing greater robustness to sub-optimal example banks.
Search
Fix author
Co-authors
- Dmitriy Bespalov 1
- Yueyan Chen 1
- Lin Lee Cheong 1
- Han Ding 1
- Haibo Ding 1
- show all...