1991
DOI: 10.1162/neco.1991.3.4.461
|View full text |Cite
|
Sign up to set email alerts
|

Neural Network Classifiers Estimate Bayesian a posteriori Probabilities

Abstract: Many neural network classifiers provide outputs which estimate Bayesian a posteriori probabilities. When the estimation is accurate, network outputs can be treated as probabilities and sum to one. Simple proofs show that Bayesian probabilities are estimated when desired network outputs are 1 of M (one output unity, all others zero) and a squared-error or cross-entropy cost function is used. Results of Monte Carlo simulations performed using multilayer perceptron (MLP) networks trained with backpropagation, rad… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

2
410
0
7

Year Published

1995
1995
2016
2016

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 933 publications
(419 citation statements)
references
References 6 publications
2
410
0
7
Order By: Relevance
“…2. For A-NNi, the simple quadratic function implies we obtain an approximation of posterior probabilities [Richard and Lippman, 1991]. This cost function choice has also proven to be efficient in the ambiguity removal phase.…”
Section: Nn-inverse Modelmentioning
confidence: 99%
“…2. For A-NNi, the simple quadratic function implies we obtain an approximation of posterior probabilities [Richard and Lippman, 1991]. This cost function choice has also proven to be efficient in the ambiguity removal phase.…”
Section: Nn-inverse Modelmentioning
confidence: 99%
“…Under these conditions a network trained with the least squares error function will approximate the conditional probability of classification. For an extended proof of this theorem the reader is directed to [14] …”
Section: Acknowledgementsmentioning
confidence: 99%
“…Many classification algorithms can be trained to generate fuzzy outputs or approximate posterior probabilities (Richard and Lippmann, 1991;Moody et al, 1996;). An important practice within the machine learning community is that classification performance can be significantly improved by simply adjusting the output threshold value instead of using a default one (e.g., 0.5) to label each target class (Provost, 2000;Zhou and Liu, 2006).…”
Section: Introductionmentioning
confidence: 99%