Tamer Ghattas


2025

pdf bib
On Pruning State-Space LLMs
Tamer Ghattas | Michael Hassid | Roy Schwartz
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing

Recent work proposed state-space models (SSMs) as an efficient alternative to transformer-based LLMs. Can these models be pruned to further reduce their computation costs? We adapt several pruning methods to the SSM structure, and apply them to four SSM-based LLMs across multiple tasks. We find that such models are quite robust to some pruning methods (e.g., WANDA), while using other methods lead to fast performance degradation.