2020
DOI: 10.48550/arxiv.2006.02601
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the Minimax Optimality of the EM Algorithm for Learning Two-Component Mixed Linear Regression

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
6
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

2
3

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 0 publications
1
6
0
Order By: Relevance
“…When the curvature around the local optimum degenerates, first-order methods such as gradient descent slow down due to vanishing gradients as the estimator gets closer to the local optimum. This phenomenon is reported in various optimization problems with degenerate landscapes in weakly separated mixture of distributions [Dwivedi et al, 2020a, Kwon et al, 2020. We can observe the same phenomenon when the rank is over-specified for low-rank matrix factorization problems.…”
Section: Related Worksupporting
confidence: 78%
See 2 more Smart Citations
“…When the curvature around the local optimum degenerates, first-order methods such as gradient descent slow down due to vanishing gradients as the estimator gets closer to the local optimum. This phenomenon is reported in various optimization problems with degenerate landscapes in weakly separated mixture of distributions [Dwivedi et al, 2020a, Kwon et al, 2020. We can observe the same phenomenon when the rank is over-specified for low-rank matrix factorization problems.…”
Section: Related Worksupporting
confidence: 78%
“…We still need to show that equation ( 5) holds (with probability at least 1 − d −c ) in this sub-linear convergence case for all iteration t with high probability. To do so, we need to use the localization technique [Kwon et al, 2020, Dwivedi et al, 2020b. Without the localization technique, the statistical error will be proportional to n −1/4 which is not tight.…”
Section: Proof Sketch For the Main Theoremmentioning
confidence: 99%
See 1 more Smart Citation
“…Recently, a seminal work [6] obtains a general framework to study the statistical guarantees of EM algorithms in the classic lowdimensional setting. In subsequent works, the convergence rates of EM algorithm under various hidden variable models are studied, including the Gaussian mixture [58,13,59,39] and mixture of linear regression [61,41,40]. Another important direction is to design variants of EM algorithms in the high-dimensional regime.…”
Section: Introductionmentioning
confidence: 99%
“…When the separation gets small (smaller δ or H), the convergence speed gets slower. This type of transition in the convergence speed of EM (the update of model parameters with Algorithm 3) is observed both in theory and practice when the overlap between mixture components gets larger (e.g.,[28]). On the other hand, the policy steadily improves regardless of the level of separation.…”
mentioning
confidence: 99%