2021
DOI: 10.48550/arxiv.2110.01602
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Clustering a Mixture of Gaussians with Unknown Covariance

Abstract: We investigate a clustering problem with data from a mixture of Gaussians that share a common but unknown, and potentially ill-conditioned, covariance matrix. We start by considering Gaussian mixtures with two equally-sized components and derive a Max-Cut integer program based on maximum likelihood estimation. We prove its solutions achieve the optimal misclassification rate when the number of samples grows linearly in the dimension, up to a logarithmic factor. However, solving the Max-cut problem appears to b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
10
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(11 citation statements)
references
References 89 publications
1
10
0
Order By: Relevance
“…Recently, [DDW21] showed that the MLE estimator for the missing labels corresponds to a Max-Cut problem, which recovers the solution when n = Ω(d). Moreover, the authors argued that while SNR drives the inherent statistical difficulty of the estimation problem, a relaxed quantity S = v 2 / Σ presumably controls the computational difficulty.…”
Section: Relation To Prior Workmentioning
confidence: 74%
See 3 more Smart Citations
“…Recently, [DDW21] showed that the MLE estimator for the missing labels corresponds to a Max-Cut problem, which recovers the solution when n = Ω(d). Moreover, the authors argued that while SNR drives the inherent statistical difficulty of the estimation problem, a relaxed quantity S = v 2 / Σ presumably controls the computational difficulty.…”
Section: Relation To Prior Workmentioning
confidence: 74%
“…In particular, this improves the state-of-the-art for this task, and provably beats all low-degree algorithms, when ρ = ω(1/ √ d). The corollary is based on the equivalence (up to rescaling) between Model 3.5 and some appropriate sub-case of Model 3.2 [DDW21]. Proof.…”
Section: Implications Of the Success Of Lllmentioning
confidence: 99%
See 2 more Smart Citations
“…Similarly, the assumption of isotropic noise is relaxable as long as the covariance tensor is known. The case of unknown covariance is much more challenging (Davis et al, 2021;Bakshi et al, 2020;Belkin and Sinha, 2010;Cai et al, 2019;Ge et al, 2015;Moitra and Valiant, 2010) even in the vector case and is beyond the scope of the current paper.…”
Section: Introductionmentioning
confidence: 98%