Proceedings of the 50th Annual ACM SIGACT Symposium on Theory of Computing 2018
DOI: 10.1145/3188745.3188748
|View full text |Cite
|
Sign up to set email alerts
|

Mixture models, robustness, and sum of squares proofs

Abstract: We use the Sum of Squares method to develop new efficient algorithms for learning wellseparated mixtures of Gaussians and robust mean estimation, both in high dimensions, that substantially improve upon the statistical guarantees achieved by previous efficient algorithms. Our contributions are:• Mixture models with separated means: We study mixtures of k distributions in d dimensions, where the means of every pair of distributions are separated by at least k ε . In the special case of spherical Gaussian mixtur… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
35
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 63 publications
(37 citation statements)
references
References 37 publications
2
35
0
Order By: Relevance
“…Can we obtain nearly-linear time robust algorithms for other inference tasks under sparsity assumptions [BDLS17] (e.g., for robust sparse mean estimation or robust sparse PCA)? Can we speed-up the convex programs obtained via the SoS hierarchy in this setting [HL18,KSS18]?…”
Section: Discussionmentioning
confidence: 99%
“…Can we obtain nearly-linear time robust algorithms for other inference tasks under sparsity assumptions [BDLS17] (e.g., for robust sparse mean estimation or robust sparse PCA)? Can we speed-up the convex programs obtained via the SoS hierarchy in this setting [HL18,KSS18]?…”
Section: Discussionmentioning
confidence: 99%
“…However, in the present literature, most of the estimation procedures with finite sample guarantees are either clustering-based, or rely on separation conditions in the analysis (e.g. [156,157,158]). Bridging this conceptual divide is one of the main motivations of the present chapter.…”
Section: Chapter 6 a Framework For Learning Mixture Modelsmentioning
confidence: 99%
“…Their technique, isotropic PCA, an affine-invariant version of PCA, is not robust (showing this is a bit more involved than for most spectral algorithms). So we turn for inspiration to the special case of spherical Gaussians for which robust algorithms have been recently discovered, with nearoptimal separation [16,11]. The key idea there is to express the identifiability of a Gaussian component in terms of a polynomial system, solve this polynomial system using a sum-of-squares semi-definite programming relaxation, and round the fractional solution obtained to a nearly correct clustering.…”
Section: Approachmentioning
confidence: 99%
“…Over the past few years, there has been significant progress in computationally efficient robust estimation, starting with mean and covariance estimation for a large family of distributions [6,17]. Such robust estimation has also been discovered for various generative models including mixtures of well-separated spherical Gaussians (early work by Brubaker [3], and improved bounds more recently [6,11,16]), Independent Component Analysis [17,16] and linear regression. In spite of impressive progress, the core motivating problem of robustly estimating a mixture of (even two) Gaussians has remained unsolved.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation