2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society 2011
DOI: 10.1109/iembs.2011.6090214
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating feature selection strategies for high dimensional, small sample size datasets

Abstract: In this work, we analyze and evaluate different strategies for comparing Feature Selection (FS) schemes on High Dimensional (HD) biomedical datasets (e.g. gene and protein expression studies) with a small sample size (SSS). Additionally, we define a new feature, Robustness, specifically for comparing the ability of an FS scheme to be invariant to changes in its training data. While classifier accuracy has been the de facto method for evaluating FS schemes, on account of the curse of dimensionality problem, it … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
20
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
8
1
1

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(20 citation statements)
references
References 12 publications
0
20
0
Order By: Relevance
“…Stepwise feature selection (27) was used with multiple linear regression analysis to select the subset of features for the classification task by using the Wilks L criterion. To reduce database bias, stepwise feature selection and classification were conducted concurrently within a leave-one-case-out cross-validation manner.…”
Section: Classifier Developmentmentioning
confidence: 99%
“…Stepwise feature selection (27) was used with multiple linear regression analysis to select the subset of features for the classification task by using the Wilks L criterion. To reduce database bias, stepwise feature selection and classification were conducted concurrently within a leave-one-case-out cross-validation manner.…”
Section: Classifier Developmentmentioning
confidence: 99%
“…Selecting a feature subset of low cardinality and high discrimination power has been a centre-stage quest since the dawn of pattern recognition [1,2,3,4]. Feature selection from high-dimensional data has been extensively studied [5,6,7,8,9,10,11,12]. In many cases, feature selection is sought as the end goal of the data analysis.…”
Section: Introductionmentioning
confidence: 99%
“…In this study, a range of methods were used. These methods are outlined below; they have been reported on extensively previously [ 30‐32,37,39,41‐59 ] and detailed methods are also included as Appendix S1.…”
Section: Methodsmentioning
confidence: 99%