1992
DOI: 10.1109/72.143378
|View full text |Cite
|
Sign up to set email alerts
|

Statistically controlled activation weight initialization (SCAWI)

Abstract: An optimum weight initialization which strongly improves the performance of the back propagation (BP) algorithm is suggested. By statistical analysis, the scale factor, R (which is proportional to the maximum magnitude of the weights), is obtained as a function of the paralyzed neuron percentage (PNP). Also, by computer simulation, the performances on the convergence speed have been related to PNP. An optimum range for R is shown to exist in order to minimize the time needed to reach the minimum of the cost fu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
39
0
1

Year Published

1994
1994
2014
2014

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 115 publications
(40 citation statements)
references
References 3 publications
0
39
0
1
Order By: Relevance
“…well known weight initialization methods; BoersK, (Boers & Kuiper, 1992), Bottou, (Bottou, 1988), KimRa, (Kim & Ra, 1991), NW, (Nguyen & Widrow, 1990), SCAWI, (Drago & Ridella, 1992), and Smieja, (Smieja, 1991).…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…well known weight initialization methods; BoersK, (Boers & Kuiper, 1992), Bottou, (Bottou, 1988), KimRa, (Kim & Ra, 1991), NW, (Nguyen & Widrow, 1990), SCAWI, (Drago & Ridella, 1992), and Smieja, (Smieja, 1991).…”
Section: Methodsmentioning
confidence: 99%
“…They concluded that the best initial weight variance is determined by the dataset, but differences for small deviations are not significant and weights in the range [−0.77, 0.77] seem to give the best mean performance. Fernández-Redondo & Hernández-Espinosa (2001) presented an extensive experimental comparison of seven weight initialization methods; those reported by Kim & Ra (1991); Li et al (1993); Palubinskas (1994); Shimodaira (1994) ;Yoon et al (1995); Drago & Ridella (1992). Researchers claim that methods described in Palubinskas (1994); Shimodaira (1994) above proved to give the better results from all methods tested.…”
Section: Random Selection Of Initial Weightsmentioning
confidence: 96%
See 1 more Smart Citation
“…Efficient weight initialization is one of the most important factors for fast convergence and generalization, and many authors have proposed various weight initialization methods. The simplest and most widely used weight initialization method is a random initialization assuming some probability distributions and some researchers proposed several modified methods to determine the best initialization interval [1][2]. Another initialization approach is to incorporate the known prior knowledge into weight initialization [3].…”
Section: Introductionmentioning
confidence: 99%
“…If we assume that there are K neurons in the hidden layer, the weight matrices 1 W and 2 W for the two-pattern class neural network can be represented by where L=((M+1)K+2(K+1)). Then, we may view the weight vector W as a weight point in an L-dimensional weight space that is defined by all the weights in the neural network.…”
mentioning
confidence: 99%