2016
DOI: 10.1080/01621459.2015.1100996
|View full text |Cite
|
Sign up to set email alerts
|

Robust Improper Maximum Likelihood: Tuning, Computation, and a Comparison With Other Methods for Robust Gaussian Clustering

Abstract: The two main topics of this article are the introduction of the "optimally tuned robust improper maximum likelihood estimator"(OTRIMLE) for robust clustering based on the multivariate Gaussian model for clusters, and a comprehensive simulation study comparing the OTRIMLE to maximum likelihood in Gaussian mixtures with and without noise component, mixtures of t-distributions, and the TCLUST approach for trimmed clustering. The OTRIMLE uses an improper constant density for modeling outliers and noise. This can b… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
61
1

Year Published

2016
2016
2024
2024

Publication Types

Select...
6
2
1
1

Relationship

2
8

Authors

Journals

citations
Cited by 61 publications
(62 citation statements)
references
References 36 publications
(45 reference statements)
0
61
1
Order By: Relevance
“…Each module of DBS was compared with various competing algorithms, and in the majority of cases, the modules outperformed those algorithms. However, the author of this work concurs with [Coretto/Hennig, 2016] that despite one's best intentions and efforts to conduct fair comparisons of various methods of visualization, projection and clustering, "ultimately it would be good to have comparisons of methods run by researchers who did not have their hand in the design of any of the methods"; this is because "(simulation) studies can always be designed that make any method 'win.' " The author also agrees with [Coretto/Hennig, 2016] that "readers need to make up their own mind about to what extent our study covered situations that are important to them."…”
Section: The Databionic Swarm (Dbs) Methodsmentioning
confidence: 81%
“…Each module of DBS was compared with various competing algorithms, and in the majority of cases, the modules outperformed those algorithms. However, the author of this work concurs with [Coretto/Hennig, 2016] that despite one's best intentions and efforts to conduct fair comparisons of various methods of visualization, projection and clustering, "ultimately it would be good to have comparisons of methods run by researchers who did not have their hand in the design of any of the methods"; this is because "(simulation) studies can always be designed that make any method 'win.' " The author also agrees with [Coretto/Hennig, 2016] that "readers need to make up their own mind about to what extent our study covered situations that are important to them."…”
Section: The Databionic Swarm (Dbs) Methodsmentioning
confidence: 81%
“…A refined estimate of the population sizes, as well as a refined identification of group membership, could be obtained by adding a confirmatory step to the tentative clustering that we obtain by our divisive procedure. The confirmatory step could also help separate small and concentrated groups of contaminated observations from background noise (Hennig & Liao, ; Coretto & Hennig, ) and highlight the relationship between our procedure and the robust fitting of mixture models. Both these topics are the subject of ongoing research.…”
Section: Discussionmentioning
confidence: 96%
“…The term noise here is inherited from the robust clustering and classification literature (see Ritter 2014), where it is understood as a "noisy cluster", that is, a cluster of points with an unstructured shape compared with the main groups in the population. The uniform distribution is a convenient choice to capture atypical group of points not having a central location, and that are scattered in density regions somewhat separated from the main bulk of the data (Banfield and Raftery 1993;Coretto and Hennig 2016;Hennig 2004).…”
Section: Cross-sectional Cluster Modelmentioning
confidence: 99%