Don’t Generate, Classify! Low-Latency Prompt Optimization with Structured Complementary Prompt

Hee-Soo Kim, Jun-Young Kim, Jeong-Hwan Lee, Seong-Jin Park, Kang-Min Kim


Abstract
Large language models (LLMs) have demonstrated strong performance across diverse natural language processing tasks. However, their performance varies significantly across different prompts, requiring careful engineering for consistent results. Manual prompt engineering requires substantial human effort and suffers from limited reproducibility. In contrast, automatic prompt optimization methods reduce manual effort but often depend on costly autoregressive generation, resulting in substantial latency overheads. To address these limitations, we present low-latency prompt optimization (LLPO), a novel framework that reframes prompt engineering as a classification problem. LLPO classifies structured prompt fields from user input through multi-task classification and populates a predefined template to generate an optimized system prompt with minimal latency. In LLM-based automatic evaluations across four question-answering benchmarks, LLPO improves answer quality by up to 26.5% in ∆win rate compared to prior automatic prompt optimization methods, while reducing latency by up to 1,956 times. Human evaluation shows that LLPO receives the highest proportion of top-ranked responses. Furthermore, we analyze the contribution of each structured prompt field to performance, highlighting the robustness of our framework.
Anthology ID:
2026.eacl-long.204
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4364–4383
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.204/
DOI:
Bibkey:
Cite (ACL):
Hee-Soo Kim, Jun-Young Kim, Jeong-Hwan Lee, Seong-Jin Park, and Kang-Min Kim. 2026. Don’t Generate, Classify! Low-Latency Prompt Optimization with Structured Complementary Prompt. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4364–4383, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Don’t Generate, Classify! Low-Latency Prompt Optimization with Structured Complementary Prompt (Kim et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.204.pdf