Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing 2018
DOI: 10.1145/3188745.3188882
|View full text |Cite
|
Sign up to set email alerts
|

Constant approximation for k-median and k-means with outliers via iterative rounding

Abstract: In this paper, we present a new iterative rounding framework for many clustering problems. Using this, we obtain an (α 1 + ≤ 7.081 + )-approximation algorithm for k-median with outliers, greatly improving upon the large implicit constant approximation ratio of Chen [16]. For k-means with outliers, we give an (α 2 + ≤ 53.002 + )-approximation, which is the first O(1)-approximation for this problem. The iterative algorithm framework is very versatile; we show how it can be used to give α 1 -and (α 1 + )approxima… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
115
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
3
2

Relationship

0
10

Authors

Journals

citations
Cited by 72 publications
(116 citation statements)
references
References 52 publications
0
115
0
1
Order By: Relevance
“…Since the power of parameterized algorithms for uncapacitated clustering is well understood, it is a natural question to understand the "capacitated VS uncapacitated question" in the FPT setting. Since clustering is a universal task, like capacitated versions, many variants of clustering tasks have been studied including k-MEDIAN/k-MEANS WITH OUTLIERS [193] and MATROID/KNAPSACK MEDIAN [194]. While no variant is proved to harder than the basic versions, it would be interesting to see whether they all have the same parameterized approximability with the basic versions.…”
Section: Capacitated Clustering and Other Variantsmentioning
confidence: 99%
“…Since the power of parameterized algorithms for uncapacitated clustering is well understood, it is a natural question to understand the "capacitated VS uncapacitated question" in the FPT setting. Since clustering is a universal task, like capacitated versions, many variants of clustering tasks have been studied including k-MEDIAN/k-MEANS WITH OUTLIERS [193] and MATROID/KNAPSACK MEDIAN [194]. While no variant is proved to harder than the basic versions, it would be interesting to see whether they all have the same parameterized approximability with the basic versions.…”
Section: Capacitated Clustering and Other Variantsmentioning
confidence: 99%
“…With this classification model (algorithm), the data objects in the same cluster become more similar compared to the data objects in the other clusters. Meanwhile, the individual centroid of each cluster and the sum of squares of distances between data objects are used to create a cost function for the minimization task that will be repeated to classify and assign every data object to a certain cluster [ 5 , 14 , 15 , 16 , 17 , 46 , 47 , 48 , 49 , 50 , 51 , 52 ]. The K-means algorithm is a clustering technique to classify input data into K clusters based on unsupervised learning.…”
Section: Related Researchmentioning
confidence: 99%
“…Gupta 等 [88] 在可以违反异 常点数量限制的条件下, 基于局部搜索技术给出了一个双标准的 O(1)-近似算法. Friggstad 等 [89] 利 用局部搜索提出了双准则 PTAS: 聚类中心有 k(1 + ϵ) 个, 针对 Euclid 空间和加倍度量空间近似比为 1 + ϵ, 针对一般度量空间近似比为 25 + ϵ. Krishnaswamy 等 [90] 给出了基于迭代线性规划舍入技术的 (53.002 + ϵ)-近似算法, 这是该问题的第一个常数近似比算法. Krishnaswamy 等 [90] 的算法思想如下: 由于带异常点 k-均值问题的自然线性规划松弛的整数间隙无界, 他们先把线性规划松弛的解舍入为 张冬梅等: k-均值问题的理论与算法综述 费用损失很少的几乎整数解, 在该解中至多有两个分数开设的中心; 由此可知, 线性规划整数间隙来 自于几乎整数解和完全整数解的间隙.…”
Section: 鲁棒 K K K-均值问题unclassified