2021
DOI: 10.48550/arxiv.2105.08810
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Sparse Spiking Gradient Descent

Abstract: There is an increasing interest in emulating Spiking Neural Networks (SNNs) on neuromorphic computing devices due to their low energy consumption. Recent advances have allowed training SNNs to a point where they start to compete with traditional Artificial Neural Networks (ANNs) in terms of accuracy, while at the same time being energy efficient when run on neuromorphic hardware. However, the process of training SNNs is still based on dense tensor operations originally developed for ANNs which do not leverage … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 35 publications
0
3
0
Order By: Relevance
“…Kheradpisheh et al use temporal backpropagation (BS4NN) on dense BSNNs (Kheradpisheh et al, 2021). We further consider several additional promising lightweight approaches to training full precision SNNs, including sparse spiking gradient descent (SSGD) which reduces overhead during gradient calculations (Perez-Nieves & Goodman, 2021), and neural heterogeneity (NH) which compresses the required number of neurons with the addition of neuronindependent parameters (Perez-Nieves et al, 2021). Interestingly, BSNNs with threshold annealing can outperform full precision SNNs in terms of accuracy using considerably less memory for model parameters.…”
Section: Comparison With Lightweight Snnsmentioning
confidence: 99%
“…Kheradpisheh et al use temporal backpropagation (BS4NN) on dense BSNNs (Kheradpisheh et al, 2021). We further consider several additional promising lightweight approaches to training full precision SNNs, including sparse spiking gradient descent (SSGD) which reduces overhead during gradient calculations (Perez-Nieves & Goodman, 2021), and neural heterogeneity (NH) which compresses the required number of neurons with the addition of neuronindependent parameters (Perez-Nieves et al, 2021). Interestingly, BSNNs with threshold annealing can outperform full precision SNNs in terms of accuracy using considerably less memory for model parameters.…”
Section: Comparison With Lightweight Snnsmentioning
confidence: 99%
“…Recent experiments have shown that rate-coded networks (at the output) are robust to sparsity-promoting regularization terms [110], [111], [113]. However, networks that rely on time-to-first-spike schemes have had less success, which is unsurprising given that temporal outputs are already sparse.…”
Section: E Activity Regularizationmentioning
confidence: 99%
“…Using pseudo-derivatives for backpropagating through the non-differential threshold function, as we use for our discrete-time EGRU, was originally proposed for feedforward spiking networks in neuromorphic hardware in [16] and developed further in [3,66]. The sparsity of learning with BPTT when using appropriate pseudo-derivatives in a discrete-time feed-forward spiking neural network was recently described in [50].…”
Section: Related Workmentioning
confidence: 99%