2015
DOI: 10.17770/etr2011vol2.973
|View full text |Cite
|
Sign up to set email alerts
|

The Choice of Metrics for Clustering Algorithms

Abstract: Methods of data analysis and automatic processing are treated as knowledge discovery. In many cases it is necessary to classify data in some way or find regularities in the data. That is why the notion of similarity is becoming more and more important in the context of intelligent data processing systems. It is frequently required to ascertain how the data are interrelated, how various data differ or agree with each other, and what the measure of their comparison is. An important part in detection of similarit… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
13
0
1

Year Published

2016
2016
2022
2022

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 25 publications
(15 citation statements)
references
References 5 publications
1
13
0
1
Order By: Relevance
“…The principal components analysis method was also excluded as this technique requires the variables to be continuous. Instead, we used the Manhattan distance/city-block metric (Grabusts 2011) to evaluate the distance between the various points (Fab-spaces). This metric was appropriate as it identifies the distance between two points as the sum of the absolute differences of all their Cartesian coordinates (i.e.…”
Section: Clustering Methodologymentioning
confidence: 99%
“…The principal components analysis method was also excluded as this technique requires the variables to be continuous. Instead, we used the Manhattan distance/city-block metric (Grabusts 2011) to evaluate the distance between the various points (Fab-spaces). This metric was appropriate as it identifies the distance between two points as the sum of the absolute differences of all their Cartesian coordinates (i.e.…”
Section: Clustering Methodologymentioning
confidence: 99%
“…1.2. CD: CD is also known as maximum value distance, 45 Lagrange, 22 and chessboard distance. 46 This distance is appropriate in cases when two objects are to be defined as different if they are different in any one dimension.…”
Section: Distance Measures Reviewmentioning
confidence: 99%
“…where p is a positive value. It is the Manhattan distance when p = 1 , and the Euclidean distance when p = 2 , whereas the Chebyshev distance is a variant of Minkowski distance where p = . This is also known as maximum value distance, 23 Lagrange, 17 and chessboard distance, 24 and can be formulated as:…”
Section: Methodsmentioning
confidence: 99%