View a PDF of the paper titled Layer-Adaptive State Pruning for Deep State Space Models, by Minseon Gwak and 3 other authors
Abstract:Due to the lack of state dimension optimization methods, deep state space models (SSMs) have sacrificed model capacity, training search space, or stability to alleviate computational costs caused by high state dimensions. In this work, we provide a structured pruning method for SSMs, Layer-Adaptive STate pruning (LAST), which reduces the state dimension of each layer in minimizing model-level output energy loss by extending modal truncation for a single system. LAST scores are evaluated using the $mathcal{H}_{infty}$ norms of subsystems and layer-wise energy normalization. The scores serve as global pruning criteria, enabling cross-layer comparison of states and layer-adaptive pruning. Across various sequence benchmarks, LAST optimizes previous SSMs, revealing the redundancy and compressibility of their state spaces. Notably, we demonstrate that, on average, pruning 33% of states still maintains performance with 0.52% accuracy loss in multi-input multi-output SSMs without retraining. Code is available at this https URL.
Submission history
From: Minseon Gwak [view email]
[v1]
Tue, 5 Nov 2024 05:50:51 UTC (448 KB)
[v2]
Tue, 14 Jan 2025 07:30:20 UTC (448 KB)
Source link
lol