2020
DOI: 10.48550/arxiv.2005.10817
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Computationally efficient sparse clustering

Abstract: We study statistical and computational limits of clustering when the means of the centres are sparse and their dimension is possibly much larger than the sample size. Our theoretical analysis focuses on the simple model, which has two clusters with centres θ and −θ.We provide a finite sample analysis of a new sparse clustering algorithm based on sparse PCA and show that it achieves the minimax optimal misclustering rate in the regime θ → ∞, matching asymptotically the Bayes error.Our results require the sparsi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 31 publications
0
1
0
Order By: Relevance
“…This framework is indeed a good benchmark, since on the one hand, it is sufficiently simple to allow us to understand the nature of the target (β 0,1 , ..., β 0,K ) -with K = 2 and β 0,1 = −β 0,2 in our bipartite framework -and to investigate the rate of convergence of estimators of the form of (2), suitably regularized by a 1 -penalty. On the other hand, the two-component high-dimensional Gaussian mixture has received recently at lot of attention [7,2,35,28,24,15,12,3,25,8,30]. Let us emphasize that our goal is not a priori to provide a state-of-the-art method, specifically designed to solve the high-dimensional…”
Section: Introductionmentioning
confidence: 99%
“…This framework is indeed a good benchmark, since on the one hand, it is sufficiently simple to allow us to understand the nature of the target (β 0,1 , ..., β 0,K ) -with K = 2 and β 0,1 = −β 0,2 in our bipartite framework -and to investigate the rate of convergence of estimators of the form of (2), suitably regularized by a 1 -penalty. On the other hand, the two-component high-dimensional Gaussian mixture has received recently at lot of attention [7,2,35,28,24,15,12,3,25,8,30]. Let us emphasize that our goal is not a priori to provide a state-of-the-art method, specifically designed to solve the high-dimensional…”
Section: Introductionmentioning
confidence: 99%