Advances in Knowledge Discovery and Data Mining
DOI: 10.1007/978-3-540-68125-0_4
|View full text |Cite
|
Sign up to set email alerts
|

Cost-Sensitive Classifier Evaluation Using Cost Curves

Abstract: The evaluation of classifier performance in a cost-sensitive setting is straightforward if the operating conditions (misclassification costs and class distributions) are fixed and known. When this is not the case, evaluation requires a method of visualizing classifier performance across the full range of possible operating conditions. This talk outlines the most important requirements for cost-sensitive classifier evaluation and introduces a technique for classifier performance visualization -the cost curve -t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Publication Types

Select...
3
2
2

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(3 citation statements)
references
References 25 publications
0
3
0
Order By: Relevance
“…In order to determine whether a scaling factor ν different from 1 would improve the discrimination of ID segments from structured proteins, we constructed ROC curves at different scaling factors. We determined an optimal cutoff value at each scaling factor by selecting the threshold with the highest Matthew's correlation coefficient (MCC) [37], [38]. Out of all the scaling factors that we tested, ν of 1 results in a classifier that has the highest AUC (Table S1).…”
Section: Resultsmentioning
confidence: 99%
“…In order to determine whether a scaling factor ν different from 1 would improve the discrimination of ID segments from structured proteins, we constructed ROC curves at different scaling factors. We determined an optimal cutoff value at each scaling factor by selecting the threshold with the highest Matthew's correlation coefficient (MCC) [37], [38]. Out of all the scaling factors that we tested, ν of 1 results in a classifier that has the highest AUC (Table S1).…”
Section: Resultsmentioning
confidence: 99%
“…Use of well-known discretization methods can cause information loss and hypotheses missing in the version space construction of a numeric data set. These discretization methods include the entropy-based discretization method [8], [9], [10], Gini-Index based method [11], Chisquare based method [12], [13], 1-rule method [14], [15], and unsupervised equal-length or equal-density methods. In fact, use of these discretization methods makes almost every version space empty as investigated by this work.…”
Section: Examplementioning
confidence: 99%
“…In other words, both classifiers perform equally well if the cost of classifying benign and malignant tumors is kept the same. However, if we would like to change the cost of classifying benign and malignant tumors, for example, we decided to give more cost for missing malignant tumors than missing benign tumors then both classifiers perform differently (see Holte & Drummond (2011)). The later observation explains why the SVM and Parzen classifier have an overlapping performance which is easy to explain from the ROC curves.…”
Section: Analyzing the Classifiers Performancementioning
confidence: 99%