@inproceedings{xu-etal-2025-procut,
title = "{P}ro{C}ut: {LLM} Prompt Compression via Attribution Estimation",
author = "Xu, Zhentao and
Li, Fengyi and
Chen, Albert C. and
Wang, Xiaofeng",
editor = "Potdar, Saloni and
Rojas-Barahona, Lina and
Montella, Sebastien",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track",
month = nov,
year = "2025",
address = "Suzhou (China)",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-industry.20/",
pages = "285--309",
ISBN = "979-8-89176-333-3",
abstract = "In large-scale industrial LLM systems, prompt templates often expand to thousands of tokens as teams iteratively incorporate sections such as task instructions, few-shot examples, and heuristic rules to enhance robustness and coverage. This expansion leads to bloated prompts that are difficult to maintain and incur significant inference latency and serving costs. To address this, we introduce Prompt Compression via Attribution Estimation (ProCut), a flexible, LLM-agnostic, training-free framework that compresses prompts through attribution analysis. ProCut segments prompt templates into semantically meaningful units, quantifies their impact on task performance, and prunes low-utility components. Through extensive experiments on five public benchmark datasets and real-world industrial prompts, we show that ProCut achieves substantial prompt size reductions (78{\%} fewer tokens in production) while maintaining or even slightly improving task performance (up to 62{\%} better than alternative methods). We further introduce an LLM-driven attribution estimator that reduces compression latency by over 50{\%}, and demonstrate that ProCut integrates seamlessly with existing prompt-optimization frameworks to produce concise, high-performing prompts."
}Markdown (Informal)
[ProCut: LLM Prompt Compression via Attribution Estimation](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-industry.20/) (Xu et al., EMNLP 2025)
ACL
- Zhentao Xu, Fengyi Li, Albert C. Chen, and Xiaofeng Wang. 2025. ProCut: LLM Prompt Compression via Attribution Estimation. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing: Industry Track, pages 285–309, Suzhou (China). Association for Computational Linguistics.