Proceedings of the Twenty-Fifth Annual Symposium on Computational Geometry 2009
DOI: 10.1145/1542362.1542419
|View full text |Cite
|
Sign up to set email alerts
|

k-means requires exponentially many iterations even in the plane

Abstract: The k-means algorithm is a well-known method for partitioning n points that lie in the d-dimensional space into k clusters. Its main features are simplicity and speed in practice. Theoretically, however, the best known upper bound on its running time (i.e. O(n kd )) can be exponential in the number of points. Recently, Arthur and Vassilvitskii [2] showed a superpolynomial worst-case analysis, improving the best known lower bound from Ω(n) to 2with a construction in d = Ω( √ n) dimensions. In [2] they also conj… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2009
2009
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 69 publications
(16 citation statements)
references
References 11 publications
0
16
0
Order By: Relevance
“…The second step identifies different classes inside the extreme outlier set. This is done by a k -means algorithm [35, 36]. The algorithm permits to classify all the elements of the outlier set in one of the k classes.…”
Section: Models and Methodsmentioning
confidence: 99%
“…The second step identifies different classes inside the extreme outlier set. This is done by a k -means algorithm [35, 36]. The algorithm permits to classify all the elements of the outlier set in one of the k classes.…”
Section: Models and Methodsmentioning
confidence: 99%
“…The k-means is a popular objective function used for clustering problems in modern data science applications, such as computer vision, machine learning, and computational geometry (Drineas et al 2004;Little and Jones 2011;Vattani 2011). It is originally proposed by Forgy (1965) and MacQueen (1967) and is often known as Lloyd's algorithm (Lloyd 1982).…”
Section: Stacking Velocity Estimation By Weighted Clusteringmentioning
confidence: 99%
“…When there truly are K clusters, and enough effort is expended, then, in some cases, K -means will converge quickly to the right solution [13,19]. On the other hand, it is known that for ill-fated configurations K -means can take a long time to converge [18]. As such, K -means must be significantly tailored and tested for use in practical applications.…”
Section: Existing Workmentioning
confidence: 99%
“…This considerably decreases the flop count of algorithms that try to minimize the above expression, as there are many fewer terms involved if K M. There are many algorithms that directly or indirectly try to minimize the above expression over the K columns Y j . However it is difficult to the find the global minimum and the quality of the local minimum may not be good, though there does not necessarily seem to be agreement over this in the literature, as the precise local minima at which the algorithm stops depends on the starting point [13,14,18].…”
Section: Introductionmentioning
confidence: 99%