2020
DOI: 10.1016/j.nic.2020.08.004
|View full text |Cite
|
Sign up to set email alerts
|

Machine Learning Algorithm Validation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
42
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
3

Relationship

1
7

Authors

Journals

citations
Cited by 82 publications
(42 citation statements)
references
References 32 publications
0
42
0
Order By: Relevance
“…Here each group has a chance to select as the testing dataset at a time and select as the training dataset at k-1 times. In general, 10-fold CV is used [23], [24]. There are several metrics to measure the accuracy of the classifier built.…”
Section: Evaluation Of the Methodologymentioning
confidence: 99%
“…Here each group has a chance to select as the testing dataset at a time and select as the training dataset at k-1 times. In general, 10-fold CV is used [23], [24]. There are several metrics to measure the accuracy of the classifier built.…”
Section: Evaluation Of the Methodologymentioning
confidence: 99%
“…While leavepair-out cross validation is considered to be a less biased approach for binary classification because it exhaustively tries every possible combination, leave-one-out cross validation is a common training-testing split in this line of research (Cohen and Pakhomov, 2020;de la Fuente Garcia et al, 2020;Luz et al, 2020). Even on very small datasets, leave-pair-out cross validation is computationally expensive (Maleki et al, 2020). In order to keep our work comparable with prior and future studies, we opted to use leave one out cross validation as the best method for maximizing the available data while reducing training bias and maintaining reproducibility (Pahikkala et al, 2008;Fraser et al, 2019;Maleki et al, 2020).…”
Section: Machine Learning Experimentsmentioning
confidence: 99%
“…Even on very small datasets, leave-pair-out cross validation is computationally expensive (Maleki et al, 2020). In order to keep our work comparable with prior and future studies, we opted to use leave one out cross validation as the best method for maximizing the available data while reducing training bias and maintaining reproducibility (Pahikkala et al, 2008;Fraser et al, 2019;Maleki et al, 2020).…”
Section: Machine Learning Experimentsmentioning
confidence: 99%
“…Finally, we used a Wilcoxon rank-sum test to assess if there was any significant difference between the performance of the models built based on scenario 2 and the performance of the models developed based on scenario 1. In order to build RF models, we followed the common practice for developing machine learning models [29]. To achieve an unbiased estimate of generalization error, 30% of the patients were randomly selected and set aside as the test group.…”
Section: Predictive Modeling Of Different Outcomes Using Machine Learningmentioning
confidence: 99%
“…Data from the remaining 70% were used for model development. The data partitioning in this paper was conducted in a stratified manner to preserve the distribution of samples for each endpoint In order to build RF models, we followed the common practice for developing machine learning models [29]. To achieve an unbiased estimate of generalization error, 30% of the patients were randomly selected and set aside as the test group.…”
Section: Predictive Modeling Of Different Outcomes Using Machine Learningmentioning
confidence: 99%