IEEE International Conference on Neural Networks 1988
DOI: 10.1109/icnn.1988.23902
|View full text |Cite
|
Sign up to set email alerts
|

Gradient descent fails to separate

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
4
0

Year Published

1989
1989
2011
2011

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(4 citation statements)
references
References 1 publication
0
4
0
Order By: Relevance
“…It is well known from the perceptron convergence theorem [2]- [4] that the weights of the perceptron will converge to a fixed point within a finite number of updates if the set of bounded training feature vectors is linearly separable, and the weights may exhibit a limit cycle behavior if the set of bounded training feature vectors is nonlinearly separable. However, the exact condition for the weights of the perceptron exhibiting the limit cycle behavior is unknown.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…It is well known from the perceptron convergence theorem [2]- [4] that the weights of the perceptron will converge to a fixed point within a finite number of updates if the set of bounded training feature vectors is linearly separable, and the weights may exhibit a limit cycle behavior if the set of bounded training feature vectors is nonlinearly separable. However, the exact condition for the weights of the perceptron exhibiting the limit cycle behavior is unknown.…”
Section: Introductionmentioning
confidence: 99%
“…B ECAUSE the implementation cost of a perceptron is low and a perceptron can classify linearly separable bounded training feature vectors [2]- [4], perceptrons [6]- [10] are widely applied in many pattern recognition systems.…”
Section: Introductionmentioning
confidence: 99%
“…This allows us to use the mean value theorem to bound d k ?y k W T X k ?W T k X k . In the following section we will consider an alternative convergence 4 The standard inequality is written as…”
mentioning
confidence: 99%
“…However, some constructive algorithms have a very limited scope and in general, they are believed to give poor generalization results. Non-constructive algorithms are much less reliable than constructive algorithms and can fail to converge even when a solution exists (Brady, Raghavan, & Slawny, 1988). Some non-constructive algorithms can even converge to 'false' solutions (Brady, Raghavan, & Slawny, 1989).…”
Section: Introductionmentioning
confidence: 99%