2017
DOI: 10.1016/j.csda.2016.08.003
|View full text |Cite
|
Sign up to set email alerts
|

A simple approach to sparse clustering

Abstract: Consider the problem of sparse clustering, where it is assumed that only a subset of the features are useful for clustering purposes. In the framework of the COSA method of Friedman and Meulman, subsequently improved in the form of the Sparse K-means method of Witten and Tibshirani, a natural and simpler hill-climbing approach is introduced. The new method is shown to be competitive with these two methods and others.

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 20 publications
(15 citation statements)
references
References 25 publications
0
15
0
Order By: Relevance
“…In addition, the correlations between GEs and CNAs are incorporated into the objective function to accommodate the GE‐regulator regulations, different from the decomposition framework. Recently, for identifying subject groups, the sparse clustering framework has attracted much attention . On one hand, with high‐dimensional omics data, differences across subject clusters are usually only associated with to a small fraction of variables.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition, the correlations between GEs and CNAs are incorporated into the objective function to accommodate the GE‐regulator regulations, different from the decomposition framework. Recently, for identifying subject groups, the sparse clustering framework has attracted much attention . On one hand, with high‐dimensional omics data, differences across subject clusters are usually only associated with to a small fraction of variables.…”
Section: Discussionmentioning
confidence: 99%
“…Recently, for identifying subject groups, the sparse clustering framework has attracted much attention. 6,7,42 On one hand, with high-dimensional omics data, differences across subject clusters are usually only associated with to a small fraction of variables. The true clustering structures may be missed if too many noninformative variables are used in clustering.…”
Section: Discussionmentioning
confidence: 99%
“…It shows that only one objects are clustered incorrectly. This diagram is different from Table 1 where we identified 11 v (with PWCMRD 0.6548) instead of 14 v (with PWC-MRD 0.6584). Table 3(b) describes the detail of the proposed method on Soybean data, all except one object is assigned correctly.…”
Section: Soybeansmall Datamentioning
confidence: 44%
“…self-organizing map network) are used for this purpose. Arias-Castro and Xiao ( [11]) proposed a sparse version of clustering method. Zhang et al ([12]) proposed a novel statistical procedure This paper devotes to find the distinctive attributes among the categorical dataset using pooled relative within-cluster mean difference, then the data is clustered upon a single distinctive attribute.…”
Section: Introductionmentioning
confidence: 99%
“…However, for high-dimensional data, typically only a (small) subset of features are responsible for clustering different groups. Based on this characteristic, many sparse clustering algorithms [4,19,31,39,44,46] have been proposed to simultaneously cluster (high-dimensional) data and select the relevant features. For a detailed overview of these methods, see e.g., Witten and Tibshirani [46] .…”
Section: Introductionmentioning
confidence: 99%