2019 International Joint Conference on Neural Networks (IJCNN) 2019
DOI: 10.1109/ijcnn.2019.8851962
|View full text |Cite
|
Sign up to set email alerts
|

Are Traditional Neural Networks Well-Calibrated?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(3 citation statements)
references
References 6 publications
0
3
0
Order By: Relevance
“…Some applications scale the output 𝑓 (đ‘„) to produce a discrete probability distribution, thus interpreting the values as "confidence" scores. However, these are difficult to calibrate [Guo et al 2017;Johansson and Gabrielsson 2019], and may be unreliable in practice. Safe ordering properties thus capture the relevant range of behaviors needed for non-relational safety, and we aim to construct neural networks that are provably safe according to such properties.…”
Section: Introductionmentioning
confidence: 99%
“…Some applications scale the output 𝑓 (đ‘„) to produce a discrete probability distribution, thus interpreting the values as "confidence" scores. However, these are difficult to calibrate [Guo et al 2017;Johansson and Gabrielsson 2019], and may be unreliable in practice. Safe ordering properties thus capture the relevant range of behaviors needed for non-relational safety, and we aim to construct neural networks that are provably safe according to such properties.…”
Section: Introductionmentioning
confidence: 99%
“…40,41 Deep neural networks are also more often poorly calibrated compared to a decade ago because of overfitting from increased application of depth, width, weight decay, and application of batch normalization techniques, which manifest in probabilistic error rather than classification error. 42,43 The posterior probabilities from the previous methods are hence often poor estimates of the actual likelihood of a positive bioactivity prediction if used directly in this way. 44,45 Despite this, the assessment of calibration performance receives little attention.…”
Section: ■ Introductionmentioning
confidence: 99%
“…Machine learning (algorithmic) behavior is a further factor that influences both the output probability range and distribution of the “raw” probability estimates generated. For example, support vector machines (SVMs) provide no direct support for probability estimates associated with every output prediction, and consequently require additional work to convert the decision function into interpretable probability estimates. , Naïve Bayes (NB) generates posterior probabilities populating extreme regions of the probability scale (very high or low values) because of repeated multiplications over conditional feature probabilities. , Conversely, random forests (RFs) bias predictions toward the midpoint when the predicted fraction of classes across the underlying trees is employed as probability estimates, and extreme values can only be achieved when an exceptionally high proportion of trees predicts either label. , Deep neural networks are also more often poorly calibrated compared to a decade ago because of overfitting from increased application of depth, width, weight decay, and application of batch normalization techniques, which manifest in probabilistic error rather than classification error. , …”
Section: Introductionmentioning
confidence: 99%