GAP: a Global Adaptive Pruning Method for Large Language Models

Zhihua Ban, Haotian Ma, Siheng Zhang, Shengyu Liu, Xichen Chen, Ming Yang


Abstract
The deployment of Large Language Models (LLMs) faces significant challenges due to high computational costs,driving the demand for effective pruning techniques. Existing structured pruning methods employ uniform compression rates across network layers, neglecting the varying importance of different network depths. To address this limitation, we propose a novel optimization framework that directly minimizes global capability loss through layer-adaptive pruning rates. The framework formulates the pruning task as a combinatorial optimization problem constrained by a total parameter budget, and an efficient dynamic programming solution is derived to determine optimal layer-wise compression rates.Experiments demonstrate that, when tuning is not included, our approach achieves comparable performance with state-of-the-art methods at high pruning rates (37-50% reduction), and shows significant advantages at low pruning rates (13-25% reduction). When tuning is included, our method achieves the best performance among the compared methods.
Anthology ID:
2025.emnlp-main.1056
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
20909–20914
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1056/
DOI:
Bibkey:
Cite (ACL):
Zhihua Ban, Haotian Ma, Siheng Zhang, Shengyu Liu, Xichen Chen, and Ming Yang. 2025. GAP: a Global Adaptive Pruning Method for Large Language Models. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 20909–20914, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
GAP: a Global Adaptive Pruning Method for Large Language Models (Ban et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1056.pdf
Checklist:
 2025.emnlp-main.1056.checklist.pdf