2021
DOI: 10.1109/tnnls.2020.3002576
|View full text |Cite
|
Sign up to set email alerts
|

A Cheap Feature Selection Approach for the K-Means Algorithm

Abstract: The increase in the number of features that need to be analyzed in a wide variety of areas, such as genome sequencing, computer vision or sensor networks, represents a challenge for the K-means algorithm. In this regard, different dimensionality reduction approaches for the K-means algorithm have been designed recently, leading to algorithms that have proved to generate competitive clusterings. Unfortunately, most of these techniques tend to have fairly high computational costs and/or might not be easy to para… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 23 publications
(9 citation statements)
references
References 37 publications
0
7
0
Order By: Relevance
“…After calculation, F(4, 12) = 2.4801 when α = 0.1, and then, we can compute that F = 8.5221 for Classifier KNN and F = 10.0225 for Classifier C4.5. It is apparent to obtain that the two values are greater than the critical value F (4,12). This result exhibits that five methods are significantly different.…”
Section: Discussionmentioning
confidence: 86%
See 1 more Smart Citation
“…After calculation, F(4, 12) = 2.4801 when α = 0.1, and then, we can compute that F = 8.5221 for Classifier KNN and F = 10.0225 for Classifier C4.5. It is apparent to obtain that the two values are greater than the critical value F (4,12). This result exhibits that five methods are significantly different.…”
Section: Discussionmentioning
confidence: 86%
“…Feature selection is an important data preprocess in the fields of granular computing and artificial intelligence [1][2][3][4][5][6]. Its main goal is to reduce redundant features and simplify the complexity of the classification model, thereby improving the generalization ability of classification model [7][8][9][10][11][12].…”
Section: Introductionmentioning
confidence: 99%
“…Laplacian scores ranks the features based on the intrinsic characteristic and distinction from other features. 40 The nearest neighbor graph G can be calculated using the nearest neighbor method, as expressed in equation (3). where G ij is the value of ij node, x i and x j is the sample i and j , m is the number of samples.…”
Section: Methodsmentioning
confidence: 99%
“…(3) The clustering algorithm based on cluster center selection represented by K-means is easily affected by the initial cluster center selection. It results in unstable clustering results and makes it easy to fall into the optimal local solution [22].…”
Section: Credit Rating Classificationmentioning
confidence: 99%