2019
DOI: 10.1002/bimj.201800293
|View full text |Cite
|
Sign up to set email alerts
|

Validation of discrete time‐to‐event prediction models in the presence of competing risks

Abstract: Clinical prediction models play a key role in risk stratification, therapy assignment and many other fields of medical decision making. Before they can enter clinical practice, their usefulness has to be demonstrated using systematic validation. Methods to assess their predictive performance have been proposed for continuous, binary, and time-to-event outcomes, but the literature on validation methods for discrete time-toevent models with competing risks is sparse. The present paper tries to fill this gap and … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
15
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 19 publications
(17 citation statements)
references
References 43 publications
0
15
0
Order By: Relevance
“…2 Distribution of c statistics across studies with 95% confidence intervals (*model was developed for more than 1 outcome or time horizon) Fig. 3 Calibration plots of (re)calibrated prognostic models to predict risk of (a) all-cause dementia and (b) Alzheimer's disease; an intercept of 0 and slope of 1 (i.e., the diagonal line) represents ideal cali-bration and more spread between the groups indicates better model performance than less spread-error bars in grouped observations represent 95% confidence intervals; Q = quartile When externally evaluating the performance of a prediction model, both discrimination and calibration should be assessed: good performance on one measure does not ensure good performance on the other [34]. Among the models that we externally validated, best performance combining both discrimination and calibration was achieved by the original model by Hogan et al [24] and the recalibrated 2009 model by Barnes et al [25] In our validation data, proxies had to be used for several predictors in both models; despite these proxies, the models obtained acceptable discrimination and good (re)calibration.…”
Section: Discussionmentioning
confidence: 99%
“…2 Distribution of c statistics across studies with 95% confidence intervals (*model was developed for more than 1 outcome or time horizon) Fig. 3 Calibration plots of (re)calibrated prognostic models to predict risk of (a) all-cause dementia and (b) Alzheimer's disease; an intercept of 0 and slope of 1 (i.e., the diagonal line) represents ideal cali-bration and more spread between the groups indicates better model performance than less spread-error bars in grouped observations represent 95% confidence intervals; Q = quartile When externally evaluating the performance of a prediction model, both discrimination and calibration should be assessed: good performance on one measure does not ensure good performance on the other [34]. Among the models that we externally validated, best performance combining both discrimination and calibration was achieved by the original model by Hogan et al [24] and the recalibrated 2009 model by Barnes et al [25] In our validation data, proxies had to be used for several predictors in both models; despite these proxies, the models obtained acceptable discrimination and good (re)calibration.…”
Section: Discussionmentioning
confidence: 99%
“…We evaluated the performance of the nomogram through discrimination and calibration in the training population and the veri cation population, respectively. Since the consistency index (C-index) is equivalent to the area under the receiver operating characteristic curve (AUC) in logistic regression, we use the AUC to evaluate the discriminative ability of the nomogram [13] . The Hosmer-Lemeshow goodness-oft test is used to evaluate the calibration of the nomogram, and a calibration curve is drawn to visualize the consistency between the predicted results and the observed results [14] .…”
Section: Discussionmentioning
confidence: 99%
“…We tested the accuracy of the nomograms by discrimination and calibration both in the primary and the validation cohort. We used the receiver operating characteristic curve (ROC) to assess the discriminative ability of the nomogram and then assessed the area under the curve (AUC) (14,15). Calibration curves were used to compare the association between actual outcomes and predicted probabilities (16).…”
Section: Discussionmentioning
confidence: 99%