Efficient Distillation of Deep Spiking Neural Networks for Full-Range Timestep Deployment

📅 2025-01-27
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the low accuracy, rigid temporal deployment (relying on fixed inference timesteps), and poor generalization across varying timesteps in Spiking Neural Networks (SNNs), this paper proposes the first universal knowledge distillation framework supporting arbitrary inference timesteps. Methodologically, it integrates spatiotemporal-aware distillation, implicit multi-timestep modeling, and joint SNN–ANN optimization. Crucially, it establishes the first theoretical guarantee of implicit model convergence across the full range of timesteps. Experiments demonstrate state-of-the-art accuracy among distilled SNNs on CIFAR-10/100, CIFAR10-DVS, and ImageNet. Moreover, the framework significantly improves energy efficiency and plug-and-play deployment flexibility—enabling seamless adaptation to diverse timesteps without retraining.

Technology Category

Application Category

📝 Abstract
Spiking Neural Networks (SNNs) are emerging as a brain-inspired alternative to traditional Artificial Neural Networks (ANNs), prized for their potential energy efficiency on neuromorphic hardware. Despite this, SNNs often suffer from accuracy degradation compared to ANNs and face deployment challenges due to fixed inference timesteps, which require retraining for adjustments, limiting operational flexibility. To address these issues, our work considers the spatio-temporal property inherent in SNNs, and proposes a novel distillation framework for deep SNNs that optimizes performance across full-range timesteps without specific retraining, enhancing both efficacy and deployment adaptability. We provide both theoretical analysis and empirical validations to illustrate that training guarantees the convergence of all implicit models across full-range timesteps. Experimental results on CIFAR-10, CIFAR-100, CIFAR10-DVS, and ImageNet demonstrate state-of-the-art performance among distillation-based SNNs training methods.
Problem

Research questions and friction points this paper is trying to address.

Simplified Spiking Neural Networks
Time-length Flexibility
Model Efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Spike Neural Networks
Adaptive Timing Framework
State-of-the-art Performance
🔎 Similar Papers
No similar papers found.
Chengting Yu
Chengting Yu
Zhejiang University
Xiaochen Zhao
Xiaochen Zhao
Research Scientist, ByteDance
computer science
L
Lei Liu
ZJU-UIUC Institute, Zhejiang University
S
Shu Yang
ZJU-UIUC Institute, Zhejiang University
Gaoang Wang
Gaoang Wang
Zhejiang University / University of Illinois Urbana-Champaign Institute
Embodied AgentComputer VisionMachine Learning
E
Erping Li
College of Information Science and Electronic Engineering, Zhejiang University; ZJU-UIUC Institute, Zhejiang University
A
Aili Wang
College of Information Science and Electronic Engineering, Zhejiang University; ZJU-UIUC Institute, Zhejiang University