Tamer Ghattas
2025
On Pruning State-Space LLMs
Tamer Ghattas
|
Michael Hassid
|
Roy Schwartz
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Recent work proposed state-space models (SSMs) as an efficient alternative to transformer-based LLMs. Can these models be pruned to further reduce their computation costs? We adapt several pruning methods to the SSM structure, and apply them to four SSM-based LLMs across multiple tasks. We find that such models are quite robust to some pruning methods (e.g., WANDA), while using other methods lead to fast performance degradation.