2022
DOI: 10.1016/j.jrras.2022.01.003
|View full text |Cite
|
Sign up to set email alerts
|

A wrapper-based feature selection approach to investigate potential biomarkers for early detection of breast cancer

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
8
0
1

Year Published

2022
2022
2025
2025

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(9 citation statements)
references
References 27 publications
0
8
0
1
Order By: Relevance
“…The result showed that the proposed model successfully classified breast cancer by employing six features, such as BMI, age, levels of glucose, MCP-1, insulin, and resistin. Alnowami et al [18] utilized three ML algorithms, i.e., DT, RF, and DT, and combined them with a sequential backward-selection model. The result showed that the optimal set of biomarkers such as levels of glucose, BMI, resistin, age, and HOMA can be utilized as features for the SVM model.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…The result showed that the proposed model successfully classified breast cancer by employing six features, such as BMI, age, levels of glucose, MCP-1, insulin, and resistin. Alnowami et al [18] utilized three ML algorithms, i.e., DT, RF, and DT, and combined them with a sequential backward-selection model. The result showed that the optimal set of biomarkers such as levels of glucose, BMI, resistin, age, and HOMA can be utilized as features for the SVM model.…”
Section: Related Workmentioning
confidence: 99%
“…Support vector machine (SVM) is an ML model that divides instances of each class from the others by locating the linear optimum hyperplane after nonlinearly mapping the original data into a high-dimensional feature space. SVMs have demonstrated superior performance for breast cancer detection as compared to conventional models [15][16][17][18]. Additionally, earlier research has demonstrated the beneficial effects of using extra-tree as the feature selection approach to increase classification accuracy in natural language processing [19], white blood cell classification [20], and network intrusion detection [21].…”
Section: Introductionmentioning
confidence: 99%
“…The characteristics were compared between Plasmodium falciparum positive and negative using the Wilcoxon Rank Sum test for continuous variables and Pearson's Chisquare test for categorical variables, with Yates's continuity correction when appropriate. (Nadeem & Jabri, 2023,Dumitrescu et al, 2022, Bayesian generalised model (Modabbernia et al, 2022;Kamau et al, 2022) decision tree (Li et al, 2022;Avanceña et al, 2022;Dasgupta et al, 2022)modelswithnested crossvalidation (Parvandeh et al, 2020;Tu et al, 2022)for parameters optimisation and a wrapperbased sequential backward features selection (Alnowami et al, 2022) Nested cross-validation. Nested cross-validations involving multiple layers of crossvalidation: inner and outer folds, were carried out on the training data set to obtain reliable classification accuracy and avoid over fitting (Parvandeh et al, 2020;Tu et al, 2022).…”
Section: Discussionmentioning
confidence: 99%
“…2.4.2 Models Development. Penalised logistic regression ( 41), ( 42), Bayesian generalised model ( 43), (44), and decision tree model ( 45), ( 46), (47) with nested cross-validation (48), (49) for parameter optimisation and wrapper-based sequential backward feature selection (50) were employed to determine the malaria type (Plasmodium falciparum positive or negative). Eighty percent of the samples (one hundred sixty samples consisting of twenty-eight and one hundred thirty-two Plasmodium falciparum positive and negative samples, respectively) were used for model training.…”
Section: Methodsmentioning
confidence: 99%
“…2.4.5 Optimal Feature Selection and Hyperparameters. Feature selection was performed using sequential backward search selection (SBSS) for each inner training set (50). The SBSS started with all features and dropped the non-informative features at each iteration, improving the model's performance.…”
Section: Datamentioning
confidence: 99%