2020
DOI: 10.48550/arxiv.2009.12703
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Adaptive EM Accelerator for Unsupervised Learning of Gaussian Mixture Models

Truong Nguyen,
Guangye Chen,
Luis Chacon

Abstract: We propose an Anderson Acceleration (AA) scheme for the adaptive Expectation-Maximization (EM) algorithm for unsupervised learning a finite mixture model from multivariate data (Figueiredo and Jain 2002). The proposed algorithm is able to determine the optimal number of mixture components autonomously, and converges to the optimal solution much faster than its non-accelerated version. The success of the AA-based algorithm stems from several developments rather than a single breakthrough (and without these, our… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 49 publications
(116 reference statements)
0
5
0
Order By: Relevance
“…To understand subtypes of suicidal ideation or suicide attempt in adult and adolescent women, we used the Gaussian Mixture Model (GMM), 30 which is a probabilistic model that assumes data were generated from the finite mixture distributions. 32 To achieve interpretable results, we only considered important features obtained through a feature selection processes during building machine learning models. By doing so, we aimed to deliver more distinguishable and easy-to-interpret cluster results.…”
Section: Methodsmentioning
confidence: 99%
“…To understand subtypes of suicidal ideation or suicide attempt in adult and adolescent women, we used the Gaussian Mixture Model (GMM), 30 which is a probabilistic model that assumes data were generated from the finite mixture distributions. 32 To achieve interpretable results, we only considered important features obtained through a feature selection processes during building machine learning models. By doing so, we aimed to deliver more distinguishable and easy-to-interpret cluster results.…”
Section: Methodsmentioning
confidence: 99%
“…We present a generalized objective function of the Expectation-Maximization (EM) algorithm below 3 . Given M data sources {D 1 , • • • , D M }, where each local data distribution D i is a mixture of K underlying distributions, the objective function of EM (Nguyen et al, 2020) maximizes the following log-likelihood function:…”
Section: A Potential Idea: Expectation-maximization (Em)mentioning
confidence: 99%
“…Assuming that unknown parameters for the entire GMM are aggregated as θ, these parameters are identified by minimizing the negative log-likelihood of data samples under a positivity constraint for the mixture weights: where β is the Lagrange multiplier [51]. Once the GMM is trained, it can be used to assign each image sample to a Gaussian component:…”
Section: ) Supervised Deep Clustering (Sdc)mentioning
confidence: 99%
“…Competing Unsupervised Methods: In the absence of label information, we demonstrated the unsupervised variant (UDC) against shallow clustering, dimensionality reduction, decomposition and deep clustering methods. We considered GMM [51] and K-Means [59] as shallow clustering baselines, principal component analysis (PCA) [60], fast independent component analysis (Fast ICA) [61] and locally linear embedding (LLE) [62] as dimensionality reduction baselines, online dictionary learning (ODL) [63] as a decomposition baseline, and a convolutional autoencoder method (CAE) [64] and traditional triplet loss as deep clustering baselines. Implementations of competing methods are described below.…”
Section: Synergic Deep Learning (Sdl)mentioning
confidence: 99%