2014
DOI: 10.1016/j.jmva.2013.11.018
|View full text |Cite
|
Sign up to set email alerts
|

Assessment of the number of components in Gaussian mixture models in the presence of multiple local maximizers

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2014
2014
2022
2022

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 45 publications
(13 citation statements)
references
References 23 publications
0
13
0
Order By: Relevance
“…However, an alternative interpretation might be that BIC was simply being parsimonious, and that flower dimensions alone might not be enough to confidently separate the species. A much more detailed look at clustering the iris data, considering many more possible modeling choices, is found in Kim and Seo (2014). However, this simple look is enough to discuss the relevant ideas.…”
Section: Introductionmentioning
confidence: 99%
“…However, an alternative interpretation might be that BIC was simply being parsimonious, and that flower dimensions alone might not be enough to confidently separate the species. A much more detailed look at clustering the iris data, considering many more possible modeling choices, is found in Kim and Seo (2014). However, this simple look is enough to discuss the relevant ideas.…”
Section: Introductionmentioning
confidence: 99%
“…Firstly, the number of Gaussian components, K, is chosen, and the G-UMM is initialised and trained on an input data set (that may or may not contain any transient faults) in the space of the raw signals or in an appropriate feature space. K should be chosen sufficiently large, such that the G-UMM can accurately describe the distribution of the data, while remaining small enough to avoid over-fitting (Kim & Seo, 2014;McKenzie & Alder, 1994) and produce physically interpretable clusters where possible to aid the setting of the parameters of the HMM in the following step. The HMM is used in order to include the temporal relationship into the fault detection scheme.…”
Section: G-umm -Hmm Based Fault Detectionmentioning
confidence: 99%
“…In our experiments, we have noticed that in low dimensions, this problem does not arise when initialized with k-means. This is a well-known phenomenon that happens because the likelihood is unbounded, and, in this case, increases by fitting the covariance matrix of a component on just one datapoint (Bishop, 2006;Kim and Seo, 2014). Intuitively, the likelihood of a GMM can increase by (a) decreasing the variance over a single datapoint or (b) extending a better fit over a large number of points.…”
Section: High Dimensional Datamentioning
confidence: 99%