2019 American Control Conference (ACC) 2019
DOI: 10.23919/acc.2019.8814665
|View full text |Cite
|
Sign up to set email alerts
|

Convergence of the Expectation-Maximization Algorithm Through Discrete-Time Lyapunov Stability Theory

Abstract: In this paper, we propose a dynamical systems perspective of the Expectation-Maximization (EM) algorithm. More precisely, we can analyze the EM algorithm as a nonlinear state-space dynamical system. The EM algorithm is widely adopted for data clustering and density estimation in statistics, control systems, and machine learning. This algorithm belongs to a large class of iterative algorithms known as proximal point methods. In particular, we re-interpret limit points of the EM algorithm and other local maximiz… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
2
2

Relationship

2
6

Authors

Journals

citations
Cited by 9 publications
(7 citation statements)
references
References 11 publications
0
7
0
Order By: Relevance
“…Our work aims further to highlight that data augmentation frameworks in settings where the amount of augmented data dominates the number of observations may lead to more accurate inferences by incorporating domain knowledge or other type of information in the augmentation (like here information regarding the geometry of the system's invariant density). Many of the algorithms employed with data augmentation frameworks exhibit only local convergence, e.g., the Expectation Maximisation algorithm employed here [81]. In settings where the initial estimate strongly deviates from its true value, naive data augmentation strategies might therefore converge to sub-optimal solutions, that do not reflect the ground truth.…”
Section: Broader Impact Statementmentioning
confidence: 99%
“…Our work aims further to highlight that data augmentation frameworks in settings where the amount of augmented data dominates the number of observations may lead to more accurate inferences by incorporating domain knowledge or other type of information in the augmentation (like here information regarding the geometry of the system's invariant density). Many of the algorithms employed with data augmentation frameworks exhibit only local convergence, e.g., the Expectation Maximisation algorithm employed here [81]. In settings where the initial estimate strongly deviates from its true value, naive data augmentation strategies might therefore converge to sub-optimal solutions, that do not reflect the ground truth.…”
Section: Broader Impact Statementmentioning
confidence: 99%
“…If x * is a fixed point of this system, using Taylor's theorem we have for some T 0 > 0, (x t+1 − x * ) ≈ g(x * )(x t − x * ) for all t ≥ T 0 . Recall that the linear rate of convergence [Romero et al, 2019] is given by, γ = lim t→∞ x t+1 − x * / x t − x * , provided the limit exists. In the above scenario, the iterates converge when g(x * ) < 1 and the rate of convergence is g(x * ).…”
Section: Asymptotic Stability Of Tangent Transform Emmentioning
confidence: 99%
“…While a significant volume of these optimization-based papers have been published at machine learning (ML) venues, only a few have been explicitly dedicated to addressing concrete ML problems, applications, or algorithms [12,13,14,15,16]. Furthermore, only a small subset of this emerging topic of research has focused directly on discrete-time analysis that is the direct result of discretizations of an underlying continuous-time version of the algorithms [3,17,18,8,19].…”
Section: Introductionmentioning
confidence: 99%