@inproceedings{yugeswardeenoo-etal-2024-question,
title = "Question-Analysis Prompting Improves {LLM} Performance in Reasoning Tasks",
author = "Yugeswardeenoo, Dharunish and
Zhu, Kevin and
O{'}Brien, Sean",
editor = "Fu, Xiyan and
Fleisig, Eve",
booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 4: Student Research Workshop)",
month = aug,
year = "2024",
address = "Bangkok, Thailand",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.acl-srw.45/",
doi = "10.18653/v1/2024.acl-srw.45",
pages = "402--413",
abstract = "Although LLMs have the potential to transform many fields, they still underperform humans in reasoning tasks. Existing methods induce the model to produce step-by-step calculations, but this research explores the question: Does making the LLM analyze the question improve its performance? We propose a novel prompting strategy called Question Analysis Prompting (QAP), in which the model is prompted to explain the question in `n' words before solving. The value of `n' influences the length of response generated by the model. QAP is evaluated on GPT-3.5 Turbo and GPT-4 Turbo on arithmetic datasets GSM8K, AQuA, and SAT and commonsense dataset StrategyQA. QAP is compared with other state-of-the-art prompts including chain-of-thought (CoT), Plan and Solve Prompting (PS+) and Take A Deep Breath (TADB). QAP outperforms all state-of-the-art prompts on AQuA and SAT datasets on both GPT-3.5 and GPT-4. QAP consistently ranks among the top-2 prompts on 75{\%} of the tests. A key factor of QAP performance can be attributed to response length, where detailed responses are beneficial when answering harder questions, but can negatively affect easy questions."
}
Markdown (Informal)
[Question-Analysis Prompting Improves LLM Performance in Reasoning Tasks](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.acl-srw.45/) (Yugeswardeenoo et al., ACL 2024)
ACL