2020
DOI: 10.7494/csci.2020.21.2.3634
|View full text |Cite
|
Sign up to set email alerts
|

Tenfold Bootstrap Procedure for Support Vector Machines

Abstract: Cross validation is often used to split input data into training and test set in Support vector machines. The two most commonly used cross validation versions are the tenfold and leave-one-out cross validation. Another commonly used resampling method is the random test/train split. The advantage of these methods is that they avoid overfitting in the model and perform model selection. They, however, can increase the computational time for fitting Support vector machines with the increase of the size of the datase… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
9
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
7
1

Relationship

1
7

Authors

Journals

citations
Cited by 12 publications
(9 citation statements)
references
References 22 publications
0
9
0
Order By: Relevance
“…Despite the adaptation of the mixed linear integer approach to classification models, it still performs slower predictions than traditional machine learning methods (Vrigazova & Ivanov, 2020b). Therefore, improving one class of methods does not guarantee the fastest classification.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Despite the adaptation of the mixed linear integer approach to classification models, it still performs slower predictions than traditional machine learning methods (Vrigazova & Ivanov, 2020b). Therefore, improving one class of methods does not guarantee the fastest classification.…”
Section: Discussionmentioning
confidence: 99%
“…The aims of this paper are first to check if the tenfold bootstrap has the computational advantage as a training/test splitting method in classification methods. Similar research was conducted for the Support Vector Machines, so this paper can be considered as an extension of (Vrigazova and Ivanov, 2020b). Secondly, to propose train/test split proportion for the bootstrap procedure to reduce computational time and preserve high accuracy of the classification model.…”
Section: Introductionmentioning
confidence: 90%
“…To gain an insight into the characteristics of the dataset and avoid bias and overfitting in machine learning algorithms, we perform 10-fold cross-validation to evaluate the predictive accuracy of all obtained models. , The whole observation data set was randomly divided into 10 parts, of which 9 parts are taken as a training set and the rest part is used for testing. This process is iterated 10 times.…”
Section: Methodsmentioning
confidence: 99%
“…For running ANOVA, PCA and the logistic regression, existing functions in Python (LogisticRegression(), sklearm.decomposition.PCA() and sklearn.Pipeline (ANOVA)) are used, while a script for running the tenfold bootstrap is created by the author. The tenfold bootstrap used in step 5 and its software realization in Python 3.6 can be found in author's previous study (Vrigazova, 2020). Right part of figure 1 illustrates the proposed algorithm.…”
Section: New Approach -The Anova Bootstrapped Pcamentioning
confidence: 99%