2018
DOI: 10.1016/j.compeleceng.2018.04.023
|View full text |Cite
|
Sign up to set email alerts
|

A fast and effective partitional clustering algorithm for large categorical datasets using a k -means based approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
38
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 57 publications
(38 citation statements)
references
References 17 publications
0
38
0
Order By: Relevance
“…Clustering methods can be classified into five types: hierarchical [2,3], partitional [4][5][6][7][8][9][10][11][12][13][14][15][16], density-based [17,18], gridbased [19] or model-based methods [20]. The aim of cluster analysis is to partition a dataset composed of N observations embedded in a d-dimensional space into k distinct clusters.…”
Section: Introductionmentioning
confidence: 99%
“…Clustering methods can be classified into five types: hierarchical [2,3], partitional [4][5][6][7][8][9][10][11][12][13][14][15][16], density-based [17,18], gridbased [19] or model-based methods [20]. The aim of cluster analysis is to partition a dataset composed of N observations embedded in a d-dimensional space into k distinct clusters.…”
Section: Introductionmentioning
confidence: 99%
“…The cost time is highly increase by processing a transactional dataset, where the quality of outcomes is affected by analyzing the numerous iterations in the process. The feature selection and dimensionality reduction techniques are developed to address this issues, where the main aim is to remove the noisy, redundant and irrelevant information by preprocessing the data [11,12]. Recently, K-means and its variants are highly used for clustering large datasets because of their higher scalability and efficiency.…”
Section: Introductionmentioning
confidence: 99%
“…The k-means has advantages, i.e. it is easy to implement grouping a large dataset, and with stable performance over different problems (Ben Salem et al, 2018 [1]; Chakraborty and Das, 2018 [2]). However, the clustering results of k-means depends on a certain number of clusters as inputs.…”
Section: Introductionmentioning
confidence: 99%