2006
DOI: 10.1007/s10115-006-0040-8
|View full text |Cite
|
Sign up to set email alerts
|

Stability of feature selection algorithms: a study on high-dimensional spaces

Abstract: With the proliferation of extremely high-dimensional

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
372
0
3

Year Published

2008
2008
2022
2022

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 535 publications
(379 citation statements)
references
References 12 publications
4
372
0
3
Order By: Relevance
“…Subsequently, feature selection is performed on each of the k subsamples, and a measure of stability or robustness is calculated. Here, following [5], we take a similarity based approach where feature stability is measured by comparing the outputs of the feature selectors on the k subsamples. The more similar all outputs are, the higher the stability measure will be.…”
Section: Estimating Stability With Instance Perturbationmentioning
confidence: 99%
See 1 more Smart Citation
“…Subsequently, feature selection is performed on each of the k subsamples, and a measure of stability or robustness is calculated. Here, following [5], we take a similarity based approach where feature stability is measured by comparing the outputs of the feature selectors on the k subsamples. The more similar all outputs are, the higher the stability measure will be.…”
Section: Estimating Stability With Instance Perturbationmentioning
confidence: 99%
“…Recent work in this area mainly focuses on the stability indices to be used for feature selection, introducing measures based on Hamming distance [4], correlation coefficients [5], consistency [6] and information theory [7]. Kalousis and coworkers also present an extensive comparative evaluation of feature selection stability over a number of high-dimensional datasets [5]. However, most of this work only focuses on the stability of single feature selection techniques, an exception being the work of [4] which describes an example combining multiple feature selection runs.…”
Section: Introductionmentioning
confidence: 99%
“…Table 6 lists one stability measure MW1 [72]. The Pearson's correlation coefficient ranges from -1 to 1.…”
Section: Weighting-score Vectormentioning
confidence: 99%
“…Several similarity measures have been proposed in the literature. Some of the older works propose using the Jaccard index sim J [6] (also referred as the Tanimoto distance) or the relative Hamming distance to define a similarity measure sim H [4]. Let us assume that we have n features in total.…”
Section: Existing Measuresmentioning
confidence: 99%