2013
DOI: 10.1007/978-3-642-40994-3_29
|View full text |Cite|
|
Sign up to set email alerts
|

Area under the Precision-Recall Curve: Point Estimates and Confidence Intervals

Abstract: Abstract. The area under the precision-recall curve (AUCPR) is a single number summary of the information in the precision-recall (PR) curve. Similar to the receiver operating characteristic curve, the PR curve has its own unique properties that make estimating its enclosed area challenging. Besides a point estimate of the area, an interval estimate is often required to express magnitude and uncertainty. In this paper we perform a computational analysis of common AUCPR estimators and their confidence intervals… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
253
0
2

Year Published

2016
2016
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 317 publications
(257 citation statements)
references
References 19 publications
2
253
0
2
Order By: Relevance
“…While the loss function was deliberately designed to weigh recall higher than precision (at β = 0.7), consistent improvements in all test performance metrics including DSC and F 2 scores on the test set indicate improved generalization through this type of training. Compared to DSC which weighs recall and precision equally, and the ROC analysis, we consider the area under the PR curves (APR, shown in Figure 2) the most reliable performance metric for such highly skewed data [8,1]. To put the work in context, we reported average DSC, F 2 , and APR scores (equal to 56.4, 57.3, and 56.0, respectively), which indicate that our approach performed very well compared to the latest results in MS lesion segmentation [6,20].…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…While the loss function was deliberately designed to weigh recall higher than precision (at β = 0.7), consistent improvements in all test performance metrics including DSC and F 2 scores on the test set indicate improved generalization through this type of training. Compared to DSC which weighs recall and precision equally, and the ROC analysis, we consider the area under the PR curves (APR, shown in Figure 2) the most reliable performance metric for such highly skewed data [8,1]. To put the work in context, we reported average DSC, F 2 , and APR scores (equal to 56.4, 57.3, and 56.0, respectively), which indicate that our approach performed very well compared to the latest results in MS lesion segmentation [6,20].…”
Section: Resultsmentioning
confidence: 99%
“…To critically evaluate the performance of the detection for the highly unbalanced (skewed) dataset, we use the Precision-Recall (PR) curve (as opposed to the receiver-operator characteristic, or ROC, curve) as well as the area under the PR curve (the APR score) [1,7,8]. For such skewed datasets, the PR curves and APR scores (on test data) are preferred figures of algorithm performance.…”
Section: Evaluation Metricsmentioning
confidence: 99%
See 1 more Smart Citation
“…We calculated precisionrecall curves by successively filtering out results based on abundances to increase precision and recalculating recall at each step, defining true and false positives in terms of the binary detection of species. The AUPR was calculated using the lower trapezoid method [69]. For subspecies, classification at varying levels complicated the analysis (e.g.…”
Section: Precision-recallmentioning
confidence: 99%
“…Each threshold is associated with a value of precision and recall, corresponding to a point in the PR space. To obtain a single performance measure from the curve, we calculate the area under the curve using a trapezoidal approximation (Boyd et al, 2013). A perfect classifier would have an 30 AUPRC of 1.0.…”
Section: Measures Estimating the Predictive Performance Of Hierarchicmentioning
confidence: 99%