2014
DOI: 10.1007/s00521-014-1730-x
|View full text |Cite
|
Sign up to set email alerts
|

Batch gradient training method with smoothing $$\boldsymbol{\ell}_{\bf 0}$$ ℓ 0 regularization for feedforward neural networks

Abstract: This paper considers the batch gradient method with the smoothing ' 0 regularization (BGSL0) for training and pruning feedforward neural networks. We show why BGSL0 can produce sparse weights, which are crucial for pruning networks. We prove both the weak convergence and strong convergence of BGSL0 under mild conditions. The decreasing monotonicity of the error functions during the training process is also obtained. Two examples are given to substantiate the theoretical analysis and to show the better sparsity… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
7
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
7
2

Relationship

0
9

Authors

Journals

citations
Cited by 19 publications
(7 citation statements)
references
References 33 publications
(36 reference statements)
0
7
0
Order By: Relevance
“…In literature, there is no standard threshold value to remove unnecessary weight connections and redundant neurons from the initially assumed neural network structure. According to [20,30], the sparsity of the learning algorithm was measured by using the number of weights whose absolute values are ≤0.0099 and ≤0.01, respectively. In this study, we have arbitrary chosen 0.00099 as a threshold value which is more less than the existing thresholds in literature.…”
Section: Hidden Neuron Selection Criterionmentioning
confidence: 99%
“…In literature, there is no standard threshold value to remove unnecessary weight connections and redundant neurons from the initially assumed neural network structure. According to [20,30], the sparsity of the learning algorithm was measured by using the number of weights whose absolute values are ≤0.0099 and ≤0.01, respectively. In this study, we have arbitrary chosen 0.00099 as a threshold value which is more less than the existing thresholds in literature.…”
Section: Hidden Neuron Selection Criterionmentioning
confidence: 99%
“…Even the L 0 , L 1 , and L 1/2 regularizers can generate sparse results, but only the single sparse weight can be selected. So, it is still a challenge for us to decide which neuron is redundant [25], [37]- [40]. As a meaningful extension of Lasso, Group Lasso mainly has to do with variable selection on groups.…”
Section: Introductionmentioning
confidence: 99%
“…e pruning method starts with a large network structure and then prunes unimportant weights or nodes [15][16][17][18][19].…”
Section: Introductionmentioning
confidence: 99%