Back to Search Start Over

Fast-SNN: Fast Spiking Neural Network by Converting Quantized ANN.

Authors :
Hu Y
Zheng Q
Jiang X
Pan G
Source :
IEEE transactions on pattern analysis and machine intelligence [IEEE Trans Pattern Anal Mach Intell] 2023 Dec; Vol. 45 (12), pp. 14546-14562. Date of Electronic Publication: 2023 Nov 03.
Publication Year :
2023

Abstract

Spiking neural networks (SNNs) have shown advantages in computation and energy efficiency over traditional artificial neural networks (ANNs) thanks to their event-driven representations. SNNs also replace weight multiplications in ANNs with additions, which are more energy-efficient and less computationally intensive. However, it remains a challenge to train deep SNNs due to the discrete spiking function. A popular approach to circumvent this challenge is ANN-to-SNN conversion. However, due to the quantization error and accumulating error, it often requires lots of time steps (high inference latency) to achieve high performance, which negates SNN's advantages. To this end, this paper proposes Fast-SNN that achieves high performance with low latency. We demonstrate the equivalent mapping between temporal quantization in SNNs and spatial quantization in ANNs, based on which the minimization of the quantization error is transferred to quantized ANN training. With the minimization of the quantization error, we show that the sequential error is the primary cause of the accumulating error, which is addressed by introducing a signed IF neuron model and a layer-wise fine-tuning mechanism. Our method achieves state-of-the-art performance and low latency on various computer vision tasks, including image classification, object detection, and semantic segmentation. Codes are available at: https://github.com/yangfan-hu/Fast-SNN.

Details

Language :
English
ISSN :
1939-3539
Volume :
45
Issue :
12
Database :
MEDLINE
Journal :
IEEE transactions on pattern analysis and machine intelligence
Publication Type :
Academic Journal
Accession number :
37721891
Full Text :
https://doi.org/10.1109/TPAMI.2023.3275769