Spiking Layer-Adaptive Magnitude-based Pruning

📅 2026-03-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitations of conventional magnitude-based pruning methods in spiking neural networks (SNNs), which neglect temporal accumulation effects, non-uniform contributions across time steps, and membrane potential stability, leading to significant performance degradation. To overcome these issues, the authors propose SLAMP, a novel framework that extends layer-adaptive pruning into the temporal domain of SNNs for the first time. By formulating a distortion-constrained optimization problem that incorporates time-aware sensitivity, SLAMP jointly allocates sparsity across both network layers and individual time steps. The method introduces a time-aware layer importance metric and naturally reduces to classical pruning strategies in the single-time-step limit, ensuring both theoretical rigor and practical applicability. Experiments on CIFAR10, CIFAR100, and CIFAR10-DVS demonstrate that SLAMP substantially reduces synaptic connections and spike operations while preserving high accuracy, thereby enhancing the deployment efficiency of SNNs.

Technology Category

Application Category

📝 Abstract
Spiking Neural Networks (SNNs) provide energy-efficient computation but their deployment is constrained by dense connectivity and high spiking operation costs. Existing magnitude-based pruning strategies, when naively applied to SNNs, fail to account for temporal accumulation, non-uniform timestep contributions, and membrane stability, often leading to severe performance degradation. This paper proposes Spiking Layer-Adaptive Magnitude-based Pruning (SLAMP), a theory-guided pruning framework that generalizes layer-adaptive magnitude pruning to temporal SNNs by explicitly controlling worst-case output distortion across layers and timesteps. SLAMP formulates sparsity allocation as a temporal distortion-constrained optimization problem, yielding time-aware layer importance scores that reduce to conventional layer-adaptive pruning in single-timestep limit. An efficient two-stage procedure is derived, combining temporal score estimation, global sparsity allocation, and magnitude pruning with retraining for stability recovery. Experiments on CIFAR10, CIFAR100, and the event-based CIFAR10-DVS datasets demonstrate that SLAMP achieves substantial connectivity and spiking operation reductions while preserving accuracy, enabling efficient and deployable SNN inference.
Problem

Research questions and friction points this paper is trying to address.

Spiking Neural Networks
magnitude-based pruning
temporal accumulation
membrane stability
spiking operation costs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Spiking Neural Networks
Magnitude-based Pruning
Layer-adaptive
Temporal Distortion
Sparsity Allocation
🔎 Similar Papers
No similar papers found.
J
Junqiao Wang
Sichuan University-Pittsburgh Institute, Sichuan University, Chengdu, 610207, China
Z
Zhehang Ye
Sichuan University-Pittsburgh Institute, Sichuan University, Chengdu, 610207, China
Yuqi Ouyang
Yuqi Ouyang
Sichuan University
Computer Vision