2017
DOI: 10.15388/informatica.2017.131
|View full text |Cite
|
Sign up to set email alerts
|

Holo-Entropy Based Categorical Data Hierarchical Clustering

Abstract: Clustering high-dimensional data is a challenging task in data mining, and clustering high-dimensional categorical data is even more challenging because it is more difficult to measure the similarity between categorical objects. Most algorithms assume feature independence when computing similarity between data objects, or make use of computationally demanding techniques such as PCA for numerical data. Hierarchical clustering algorithms are often based on similarity measures computed on a common feature space, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 50 publications
0
6
0
Order By: Relevance
“…The detailed analysis and classification are sequentially presented in Section 3.2. Specifically, there are five studies related to hierarchical clustering, with four of them based on rough-set theory (MTMDP [38], MGR [41], MNIG [47], and HPCCD [96]), except the P-ROCK [106]. Two studies focus on agglomerative hierarchical clustering, while the remaining three focus on divisive hierarchical clustering.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…The detailed analysis and classification are sequentially presented in Section 3.2. Specifically, there are five studies related to hierarchical clustering, with four of them based on rough-set theory (MTMDP [38], MGR [41], MNIG [47], and HPCCD [96]), except the P-ROCK [106]. Two studies focus on agglomerative hierarchical clustering, while the remaining three focus on divisive hierarchical clustering.…”
Section: Discussionmentioning
confidence: 99%
“…Furthermore, there has been significant advancement in the performance of previous algorithms, shown by the improvement of the min-min-roughness (MMR) algorithm [99]. Divisive, based on an information-theoretic approach MMR, MGR, MDA [103], TR [104] Sun et al ( 2017) HPCCD [96] Agglomerative, based on an information-theoretic approach MGR [41], COOLCAT, LIMBO [105], K-modes,…”
Section: Hierarchical Clusteringmentioning
confidence: 99%
See 1 more Smart Citation
“…In information theory, the measure of uncertainty in a random variable is described by Shannon's entropy in equation in (9). The concept of entropy is always used to represent disorder, chaos, or unpredictability in a dataset [12]. Hence, entropy in the proposed GrC compatibility measure characterizes the reluctance of granules to be merged.…”
Section: Use Of Information Theory To Measure Uncertaintymentioning
confidence: 99%
“…For this aim the following Shannon entropy (Shannon, 1948) function can be used. This function previously has been used as the objective function of transportation problems (Ojha et al, 2009) and other optimization problems (Sun et al, 2017).…”
Section: Non-linear Formulationmentioning
confidence: 99%