2018
DOI: 10.1088/1742-6596/1007/1/012007
|View full text |Cite
|
Sign up to set email alerts
|

Attribute Weighting Based K-Nearest Neighbor Using Gain Ratio

Abstract: Abstract. K-Nearest Neighbor (KNN) is a good classifier, but fro m several studies, the result performance accuracy of KNN still lo wer than other methods. One of the causes of the low accuracy produced, because each attribute has the same effect on the classification process, while some less relevant characteristics lead to mis s-classification of the class assignment for new data. In this research, we proposedAttribute Weighting Based K-Nearest Neighbor Using Gain Rat io as a parameter to see the correlation… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0
4

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(11 citation statements)
references
References 7 publications
0
7
0
4
Order By: Relevance
“…4. Hayes Roth dataset [12] The dataset consists 160 instances, five attributes and three classes on categorization of human subjects.…”
Section: Datasetsmentioning
confidence: 99%
See 1 more Smart Citation
“…4. Hayes Roth dataset [12] The dataset consists 160 instances, five attributes and three classes on categorization of human subjects.…”
Section: Datasetsmentioning
confidence: 99%
“…For F1score (Table 4) except Gini-index for Hayes-Roth dataset (0.81), the proposed measure outperforms IG and Gini: 0.88 for zoo, 0.97 for Breast, 0.79 for Car evaluation. In Height Table 5, proposed is equal to IG and Gini-index for Hayes-Roth (8) and Car Evaluation (12). However, for zoo (6) and breast dataset (6), IG is better.…”
Section: Performancementioning
confidence: 99%
“…For each feature, the minimum value of the feature is changed to 0, the maximum value is changed to 1, and every other value is converted to a decimal between 0 and 1. The application of this method can be seen in the model equation 7 as follows [13]:…”
Section: K-nearest Neighbor (Knn) Classificationmentioning
confidence: 99%
“…Since this technique is directly geared towards maximizing the fisher score, while the mean and variance of each class can be easily calculated, it has the potential to be applied to multi-class data. Various feature quality test metrics, such as Information Gain [28]- [30], Gain ratio [31], Gini Decrease [32], Anova [33], [34], Chi-Square [35], ReliefF [36], [37], and Fast Correlation-Based Feature selection (FCBF) [38], [39], are used to test the feature quality of the proposed Box-Cox transformation and Quadratic transformation, before finally comparing their respective performance.…”
Section: Introductionmentioning
confidence: 99%