GRAD: Generative Retrieval-Aligned Demonstration Sampler for Efficient Few-Shot Reasoning

Oussama Gabouj, Kamel Charaf, Ivan Zakazov, Nicolas Baldwin, Robert West


Abstract
Large Language Models (LLMs) achieve strong performance across diverse tasks, but their effectiveness often depends on the quality of the provided context. Retrieval-Augmented Generation (RAG) enriches prompts with external information, but its reliance on static databases constrains adaptability and can result in irrelevant demonstrations. In this work, we propose a Generative Retrieval-Aligned Demonstrator (GRAD), a dynamic demonstration-based approach where an LLM model is trained to generate input-specific concise demonstrations. By tailoring demonstrations to each input, our method offers better contextual support than traditional RAG approaches. We demonstrate the superiority of GRAD under budget constraints, where we limit both the number of tokens used per demonstration and the number of tokens used for the final output. Trained solely on a math dataset, GRAD consistently outperforms strong baselines on Qwen2.5-14B across mathematical reasoning and advanced STEM questions, highlighting GRAD’s robust generalization to out-of-distribution (OOD) domains such as physics, chemistry, and computer science. Furthermore, we show that demonstrations generated by trained smaller models can effectively guide larger target models, reducing training costs while maintaining competitive accuracy. Overall, this work introduces a scalable demonstration generator model presenting the first step toward a dynamic few-shot learning paradigm in resource-constrained settings. We release the code used for the project: https://github.com/charafkamel/GRAD-demonstration-sampler
Anthology ID:
2025.findings-emnlp.1047
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
19226–19244
Language:
URL:
https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.1047/
DOI:
10.18653/v1/2025.findings-emnlp.1047
Bibkey:
Cite (ACL):
Oussama Gabouj, Kamel Charaf, Ivan Zakazov, Nicolas Baldwin, and Robert West. 2025. GRAD: Generative Retrieval-Aligned Demonstration Sampler for Efficient Few-Shot Reasoning. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 19226–19244, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
GRAD: Generative Retrieval-Aligned Demonstration Sampler for Efficient Few-Shot Reasoning (Gabouj et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-luhme/2025.findings-emnlp.1047.pdf
Checklist:
 2025.findings-emnlp.1047.checklist.pdf