1999
DOI: 10.1109/72.774274
|View full text |Cite
|
Sign up to set email alerts
|

A local minimum for the 2-3-1 XOR network

Abstract: It was assumed proven that two-layer feedforward neural networks with t-1 hidden nodes, when presented with t input patterns, can not have any suboptimal local minima on the error surface. In this paper, however, we shall give a counterexample to this assumption. This counterexample consists of a region of local minima with nonzero error on the error surface of a neural network with three hidden nodes when presented with four patterns (the XOR problem). We will also show that the original proof is valid only w… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
10
0

Year Published

2001
2001
2020
2020

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 8 publications
0
10
0
Order By: Relevance
“…Results in [18,19] claim that, with the presence of dummy units, only T −1 units in the hidden layer are required to achieve exact learning and eliminate all suboptimal local minima. Such a statement has been shown to be false by counterexample utilising the XOR problem [16]. In the following proposition, we reinvestigate this problem as a concrete example of applying Proposition 3.…”
Section: Mlps With One Hidden Layermentioning
confidence: 95%
See 3 more Smart Citations
“…Results in [18,19] claim that, with the presence of dummy units, only T −1 units in the hidden layer are required to achieve exact learning and eliminate all suboptimal local minima. Such a statement has been shown to be false by counterexample utilising the XOR problem [16]. In the following proposition, we reinvestigate this problem as a concrete example of applying Proposition 3.…”
Section: Mlps With One Hidden Layermentioning
confidence: 95%
“…Firstly, we revisit some classic results on MLPs with only one hidden layer [18,19], and exemplify our analysis using the classic XOR problem [15,12,13,16]. For a learning task with T unique training samples, a finite exact approximator is realisable with a two-layer MLP (L = 2) having T units in the hidden layer, and its training process exempts from suboptimal local minima.…”
Section: Mlps With One Hidden Layermentioning
confidence: 99%
See 2 more Smart Citations
“…However, with only two middle nodes present, the network can get stuck in a local minimum while trying to navigate the error surface. When three nodes are present this is much less likely, which explains the increased performance from adding a third neuron to the middle layer (Sprinkhuizen-Kuyper & Boers, 1999). It appears that the MSSW performs as well as the other heuristics.…”
Section: Xor Problemmentioning
confidence: 99%