2023
DOI: 10.3390/s23136077
|View full text |Cite
|
Sign up to set email alerts
|

k-Fold Cross-Validation Can Significantly Over-Estimate True Classification Accuracy in Common EEG-Based Passive BCI Experimental Designs: An Empirical Investigation

Abstract: In passive BCI studies, a common approach is to collect data from mental states of interest during relatively long trials and divide these trials into shorter “epochs” to serve as individual samples in classification. While it is known that using k-fold cross-validation (CV) in this scenario can result in unreliable estimates of mental state separability (due to autocorrelation in the samples derived from the same trial), k-fold CV is still commonly used and reported in passive BCI studies. What is not known i… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 11 publications
(6 citation statements)
references
References 17 publications
0
6
0
Order By: Relevance
“…Also, the results of fve-fold cross-validation protocol are given Table 13. As observed, this protocol outperforms the cross-subject cross-validation, while due to the following reasons, the cross-subject protocol is most popular than the k-fold one in BCI applications [32][33][34].…”
Section: Efect Of Data Augmentationmentioning
confidence: 77%
“…Also, the results of fve-fold cross-validation protocol are given Table 13. As observed, this protocol outperforms the cross-subject cross-validation, while due to the following reasons, the cross-subject protocol is most popular than the k-fold one in BCI applications [32][33][34].…”
Section: Efect Of Data Augmentationmentioning
confidence: 77%
“…However, in machine learning models based on radiomics, k-fold cross-validation risks data leakage and introduces errors. Also, inappropriate k values, imbalanced training and validation set division, and unshuffled data order affect model prediction [ 59 , 60 ]. This study adopted random grouping and k-fold cross-validation for parameter selection during model building.…”
Section: Discussionmentioning
confidence: 99%
“…For model evaluation, we employ k-fold cross-validation, indicated by the cv = 5 parameter in the RandomizedSearchCV function. By specifying cv = 5, we perform 5-fold cross-validation, iteratively splitting the training data into five equal-sized folds for training and validation [37]. Utilizing multiple folds enhances performance estimates' robustness and reduces variability compared to a single train-test split.…”
Section: Model Training and Evaluationmentioning
confidence: 99%