2016
DOI: 10.3390/e18020059
|View full text |Cite
|
Sign up to set email alerts
|

An Optimization Approach of Deriving Bounds between Entropy and Error from Joint Distribution: Case Study for Binary Classifications

Abstract: Abstract:In this work, we propose a new approach of deriving the bounds between entropy and error from a joint distribution through an optimization means. The specific case study is given on binary classifications. Two basic types of classification errors are investigated, namely, the Bayesian and non-Bayesian errors. The consideration of non-Bayesian errors is due to the facts that most classifiers result in non-Bayesian solutions. For both types of errors, we derive the closed-form relations between each bou… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(6 citation statements)
references
References 40 publications
0
6
0
Order By: Relevance
“…where An upper bound on binary classification errors P e , which is tighter than Kovalevskij's inequality, has been reported recently [13],…”
Section: Bounds On P Ementioning
confidence: 96%
See 3 more Smart Citations
“…where An upper bound on binary classification errors P e , which is tighter than Kovalevskij's inequality, has been reported recently [13],…”
Section: Bounds On P Ementioning
confidence: 96%
“…Shannon mutual information, I S , has long been used as a metric to quantify the task-specific fidelity of a measurement with respect to classification tasks [8]. This is because I S is related to the error probability, P e through Fano's inequality [10] and Kovalevskij's inequality [11][12][13].…”
Section: Tsi For Binary Classification Tasksmentioning
confidence: 99%
See 2 more Smart Citations
“…The proof is neglected in this paper, but it can be given based on the study of bounds between entropy and error (cf. [27] and references therein). The significance of Theorem 1 implies that an optimization of information measure may not guarantee to achieve an optimization of the empirically-defined similarity measure.…”
Section: (Dis)similarity Measures In Machine Learningmentioning
confidence: 99%