2021
DOI: 10.1016/j.jocs.2021.101445
|View full text |Cite
|
Sign up to set email alerts
|

K and starting means for k-means algorithm

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 41 publications
(15 citation statements)
references
References 28 publications
0
15
0
Order By: Relevance
“…Based on these components, hierarchical clustering was placed. Ward's method was applied to clarify the ideal number of clusters; however, the classification was done by the K-means method since the hierarchical clustering instead used "complementary" rather than non-hierarchical methods [38]. The statistical test identified one outlier, but the falsely highlighted shop was just an extreme point since nearly all the examined attributes were available in its case.…”
Section: Resultsmentioning
confidence: 99%
“…Based on these components, hierarchical clustering was placed. Ward's method was applied to clarify the ideal number of clusters; however, the classification was done by the K-means method since the hierarchical clustering instead used "complementary" rather than non-hierarchical methods [38]. The statistical test identified one outlier, but the falsely highlighted shop was just an extreme point since nearly all the examined attributes were available in its case.…”
Section: Resultsmentioning
confidence: 99%
“…Notice that, even though the hostility measure uses k-means as the base of the method, it is not affected by its main drawbacks. The k-means method depends on the initialization and cannot form non-convex shapes [9]. Nevertheless, the hostility measure overcomes these problems thanks to its initialization with a high value for k to maximize the number of layers and, consequently, the resulting information to combine.…”
Section: Proposed Methodsmentioning
confidence: 99%
“…While among the partitional clustering algorithms, the KMA is currently the most popular one, and its implementation consists of five main steps: (1) specify the NC m; (2) select m samples randomly as the ICCs; (3) classify each sample into the cluster where its nearest center is located; (4) update the clustering centers by treating the mean of each cluster as a new center; and (5) repeat steps (3) and (4) until the clustering centers no longer change [26]. In a subsequent study, Sculley [27] improved the KMA and proposed the Mini-batch K-means (MBK) to accelerate the clustering speed by randomly selecting a subset instead of the whole dataset to train the ICCs, but like KMA, its clustering performance is unstable and very sensitive to the ICCs [28]. A representative partition-based clustering algorithm K-means++ proposed by Arthur and Vassilvitskii provided such an effective solution to determine the ICCs, as shown in Algorithm 1, in which the ICCs are determined based on a D 2 weighting method following the principle that the larger the distances among the ICCs, the more reasonable the selection of the ICCs, and it can effectively reduce the possibility of multiple ICCs appearing in the same cluster [29].…”
Section: Clustering Algorithmsmentioning
confidence: 99%