2023
DOI: 10.18517/ijaseit.13.1.16706
|View full text |Cite
|
Sign up to set email alerts
|

A Dataset-Driven Parameter Tuning Approach for Enhanced K-Nearest Neighbour Algorithm Performance

Abstract: The number of Neighbours (k) and distance measure (DM) are widely modified for improved kNN performance. This work investigates the joint effect of these parameters in conjunction with dataset characteristics (DC) on kNN performance. Euclidean; Chebychev; Manhattan; Minkowski; and Filtered distances, eleven k values, and four DC, were systematically selected for the parameter tuning experiments. Each experiment had 20 iterations, 10-fold cross-validation method and thirty-three randomly selected datasets from … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
2
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 0 publications
0
2
0
Order By: Relevance
“…SVM, for instance, is renowned for its ability to handle high-dimensional data and its robustness to outliers, but it may suffer from sensitivity to kernel function selection and parameter tuning [ 29 ]. KNN, on the other hand, is a simple yet effective method for classification and regression, but its performance can be heavily influenced by the choice of the number of neighbors and the distance metric [ 30 ]. RF, as an ensemble method, often achieves good performance by combining the predictions of multiple decision trees.…”
Section: Discussionmentioning
confidence: 99%
“…SVM, for instance, is renowned for its ability to handle high-dimensional data and its robustness to outliers, but it may suffer from sensitivity to kernel function selection and parameter tuning [ 29 ]. KNN, on the other hand, is a simple yet effective method for classification and regression, but its performance can be heavily influenced by the choice of the number of neighbors and the distance metric [ 30 ]. RF, as an ensemble method, often achieves good performance by combining the predictions of multiple decision trees.…”
Section: Discussionmentioning
confidence: 99%
“…As in the forward propagation Equation 9 and backpropagation formulae, the weights and biases are computed and modified at the k + 1st step (Equation 10). The error produced by the network, based on the percentage absolute error (APE) and the average percentage absolute error, is the condition for ending learning for the training process (MAPE) [27]. Following training, we move on to evaluate the network's error in order to select the most effective network for forecasting.…”
Section: Methodsmentioning
confidence: 99%
“…To predict product defects using data collected from the factory that had undergone preprocessing, a total of five machine learning-based classification models were utilized: K-Nearest Neighbors Classifier, Decision Trees Classifier, Random Forest Classifier, Extra Trees Classifier, and Gradient Boosting Classifier. In employing each model, parameters such as the number of neighbors (n_neighbors), the number of trees (n_estimators), the maximum depth of the trees (max_depth), and the learning rate (learning_rate) were meticulously controlled to ensure a fair comparison of accuracy across models [32][33][34][35][36]. Furthermore, the Grid Search method was employed to identify the most optimal combination of parameters for each model [37].…”
Section: Anomaly Detection Using Machine Learning Modelsmentioning
confidence: 99%