2014
DOI: 10.1016/j.csda.2012.12.008
|View full text |Cite
|
Sign up to set email alerts
|

Model-based clustering of high-dimensional data: A review

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
232
0

Year Published

2015
2015
2020
2020

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 385 publications
(252 citation statements)
references
References 95 publications
2
232
0
Order By: Relevance
“…This is achieved by an iterative process for determining optimal clustering centers and in this iterative process similarities are calculated based on the square error criterion [8]. Clustering methods have * Correspondence: omerfaruk.ertugrul@batman.edu.tr been employed in many research problems such as biomedical datasets [11], high-dimensional problems [12], and times signals [13]. In addition to clustering, these methods were also employed to categorize samples in order to reduce the length of the dataset and determine ideal exemplars as a category definition [14][15][16].…”
Section: Introductionmentioning
confidence: 99%
“…This is achieved by an iterative process for determining optimal clustering centers and in this iterative process similarities are calculated based on the square error criterion [8]. Clustering methods have * Correspondence: omerfaruk.ertugrul@batman.edu.tr been employed in many research problems such as biomedical datasets [11], high-dimensional problems [12], and times signals [13]. In addition to clustering, these methods were also employed to categorize samples in order to reduce the length of the dataset and determine ideal exemplars as a category definition [14][15][16].…”
Section: Introductionmentioning
confidence: 99%
“…The analysis of high dimensional data implies to bear in mind data whose dimension is larger than dimensions considered in classical multivariate analysis. As indicated by Bouveyron et al [43], when conventional methods deal with high dimensional data they are susceptible to suffer the well-known curse of dimensionality, where considering a large number of irrelevant, redundant and noisy attributes leads to important prediction errors. Hence operate with this data implies the need for more specific and complex algorithms.…”
Section: Design Principlesmentioning
confidence: 99%
“…One may speed up the EM-BIC scheme by placing reasonable bounds on M or by using other efficient fitting schemes (e.g., Sondergaard and Lermusiaux 2013a;Bouveyron and Brunet-Saumard 2014). Convergence can also be accelerated by choosing a suitable initial guess for the unknown mixture components in the EM algorithm.…”
Section: ) Computational and Storage Costsmentioning
confidence: 99%