2011
DOI: 10.1093/bib/bbr008
|View full text |Cite
|
Sign up to set email alerts
|

Caveats and pitfalls of ROC analysis in clinical microarray research (and how to avoid them)

Abstract: The receiver operating characteristic (ROC) has emerged as the gold standard for assessing and comparing the performance of classifiers in a wide range of disciplines including the life sciences. ROC curves are frequently summarized in a single scalar, the area under the curve (AUC). This article discusses the caveats and pitfalls of ROC analysis in clinical microarray research, particularly in relation to (i) the interpretation of AUC (especially a value close to 0.5); (ii) model comparisons based on AUC; (ii… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
84
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 115 publications
(85 citation statements)
references
References 51 publications
1
84
0
Order By: Relevance
“…But this choice is acceptable, in clinical practice, only when SE or SP have high values; the informational approach and the figure 4 explain us why. (ii) from the classifier point of view, the best reasonable choice is to use the decision threshold which maximize the summation T P + T N [3], or something more elaborated, such as any objective function that is a linear combination of true and false positive rates via the convex hull [8]. (iii) another possible method is maximizing SE + SP , that is equivalent to the use of the so-called Youden index [9]; it uses the maximum vertical distance of ROC curve from the point (x, y) on the diagonal (chance) line.…”
Section: Appendixmentioning
confidence: 99%
See 3 more Smart Citations
“…But this choice is acceptable, in clinical practice, only when SE or SP have high values; the informational approach and the figure 4 explain us why. (ii) from the classifier point of view, the best reasonable choice is to use the decision threshold which maximize the summation T P + T N [3], or something more elaborated, such as any objective function that is a linear combination of true and false positive rates via the convex hull [8]. (iii) another possible method is maximizing SE + SP , that is equivalent to the use of the so-called Youden index [9]; it uses the maximum vertical distance of ROC curve from the point (x, y) on the diagonal (chance) line.…”
Section: Appendixmentioning
confidence: 99%
“…Fallacy of the undistributed middle -ROC curves, even if widely used, have at least one significant pitfall, that is the so called "fallacy of the undistributed middle" [8]: all random models score an AUC of 0.5, but not every model that scores an AUC of 0.5 is random; in other words AU C = 0.5 does not necessarily imply that the classifier is no better than random guessing. An interesting example of this situation is given in figure 4a of [3], where the "twothresholds classifier" assumes that a certain quantity q, being below a threshold t 1 or exceeding a threshold t 2 > t 1 , indicates disease; while q ∈ [t 1 , t 2 ] means normal. Systolic blood pressure or expression of a gene could be an example of such a situation.…”
Section: Appendixmentioning
confidence: 99%
See 2 more Smart Citations
“…It has been used for model selection in various applications, ranging from data mining competitions to biomedical tests (Berrar & Flach, 2012). The AUC is the area under the ROC curve, which depicts the tradeoffs between the false positive rate (or 1 minus specificity, depicted on the x-axis) and the true positive rate (or sensitivity, depicted on the y-axis).…”
Section: Area Under the Roc Curve (Auc)mentioning
confidence: 99%