2012
DOI: 10.1016/j.patcog.2011.12.008
|View full text |Cite
|
Sign up to set email alerts
|

An unsupervised approach to feature discretization and selection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
66
0

Year Published

2012
2012
2022
2022

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 123 publications
(66 citation statements)
references
References 20 publications
0
66
0
Order By: Relevance
“…Therefore, the methods using this approach are typically fast but they need a threshold as the stopping criterion for feature selection. Several filter-based methods have been proposed in the literature including information gain, 10 gain ratio, 11 term variance, 12 Gini index, 13 Laplacian score, 14 Fisher score, 15 minimal-redundancy-maximal-relevance, 16 random subspace method, 17 relevance-redundancy feature selection, 18 unsupervised feature selection method based on ant colony optimization (UFSACO), 19 relevance-redundancy feature selection based on ant colony optimization (RRFSACO), 20 graph clustering with node centrality for feature selection method (GCNC), 21 and graph clustering based ant colony optimization feature selection method (GCACO). 22 Wrapper-based methods combine feature selection with the design of the classifier and evaluate the feature subsets on the basis of the accuracy of classification.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Therefore, the methods using this approach are typically fast but they need a threshold as the stopping criterion for feature selection. Several filter-based methods have been proposed in the literature including information gain, 10 gain ratio, 11 term variance, 12 Gini index, 13 Laplacian score, 14 Fisher score, 15 minimal-redundancy-maximal-relevance, 16 random subspace method, 17 relevance-redundancy feature selection, 18 unsupervised feature selection method based on ant colony optimization (UFSACO), 19 relevance-redundancy feature selection based on ant colony optimization (RRFSACO), 20 graph clustering with node centrality for feature selection method (GCNC), 21 and graph clustering based ant colony optimization feature selection method (GCACO). 22 Wrapper-based methods combine feature selection with the design of the classifier and evaluate the feature subsets on the basis of the accuracy of classification.…”
Section: Related Workmentioning
confidence: 99%
“…In the case of S-shaped and V-shaped conversion functions, the number of each dimension feature data belonging to the optimal feature subset is counted, as shown in Figure 7. It can be seen from Figure 7 that the number of the 3rd, 13th, 22nd, 23rd, 24th, and 31st dimensions of the optimal feature subsets is the largest, but the final optimal feature subsets {1, 3,10,11,12,13,14,18,19,21,23,24,29,30,31} do not incorporate all of these higherfrequency features (that is, the optimal feature subset is not a simple combination of features with high frequency). The optimal feature subset does not necessarily include features with high frequency, since the feature subset selected by features of high frequency only may not have the best classification effect.…”
Section: V1mentioning
confidence: 99%
“…Not all metrics cater to all the varieties as an example the most common metrics indicating the strength of the linear relationship between two variables; Pearson product-moment correlation coefficient needs both the variables to be numeric. Many algorithms require the features to be discrete There is an elaborate note in [42], on use of feature discretization and feature selection in conjunction. Many algorithms deal only with discretized data.…”
Section: Univariate Multivariate All Types Of Datamentioning
confidence: 99%
“…Unsupervised Methods Recently, two scalar unsupervised FD methods, based on the Linde-Buzo-Gray (LBG) algorithm [25], have been proposed [26]. The first method, named U-LBG1, applies the LBG algorithm individually to each feature, and stops when the mean square error (MSE) distortion falls below a threshold ∆ or when a maximum number, q, of bits per feature is reached.…”
Section: Feature Discretizationmentioning
confidence: 99%
“…. , 10} has been shown to be adequate [26] for different types of data. Naturally, U-LBG1 may discretize features using a variable number of bits.…”
Section: Feature Discretizationmentioning
confidence: 99%