@inproceedings{zhao-etal-2025-dac,
    title = "{DAC}: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression",
    author = "Zhao, Yi  and
      Li, Zuchao  and
      Zhao, Hai  and
      Qi, Baoyuan  and
      Guoming, Liu",
    editor = "Che, Wanxiang  and
      Nabende, Joyce  and
      Shutova, Ekaterina  and
      Pilehvar, Mohammad Taher",
    booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    month = jul,
    year = "2025",
    address = "Vienna, Austria",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.952/",
    doi = "10.18653/v1/2025.acl-long.952",
    pages = "19395--19407",
    ISBN = "979-8-89176-251-0",
    abstract = "Task-agnostic prompt compression leverages the redundancy in natural language to reduce computational overhead and enhance information density within prompts, especially in long-context scenarios. Existing methods predominantly rely on information entropy as the metric to compress lexical units, aiming to achieve minimal information loss. However, these approaches overlook two critical aspects: (i) the importance of attention-critical tokens at the algorithmic level, and (ii) shifts in information entropy during the compression process. Motivated by these challenges, we propose a dynamic attention-aware approach for task-agnostic prompt compression (DAC). This approach effectively integrates entropy and attention information, dynamically sensing entropy shifts during compression to achieve fine-grained prompt compression. Extensive experiments across various domains, including LongBench, GSM8K, and BBH, show that DAC consistently yields robust and substantial improvements across a diverse range of tasks and LLMs, offering compelling evidence of its efficacy."
}Markdown (Informal)
[DAC: A Dynamic Attention-aware Approach for Task-Agnostic Prompt Compression](https://preview.aclanthology.org/ingest-emnlp/2025.acl-long.952/) (Zhao et al., ACL 2025)
ACL