2013 International Conference on Computational and Information Sciences 2013
DOI: 10.1109/iccis.2013.261
|View full text |Cite
|
Sign up to set email alerts
|

An Improved Density-Based Method for Reducing Training Data in KNN

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
3
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 2 publications
0
3
0
Order By: Relevance
“…For the KNN algorithm, researchers have made many improvements. In 2013, Zhu et al [6] proposed an improvement method based on density. is method reduces the training data and the computational cost of the KNN algorithm by a way of merging.…”
Section: Related Workmentioning
confidence: 99%
“…For the KNN algorithm, researchers have made many improvements. In 2013, Zhu et al [6] proposed an improvement method based on density. is method reduces the training data and the computational cost of the KNN algorithm by a way of merging.…”
Section: Related Workmentioning
confidence: 99%
“…And many scholars, at home or from abroad, have proposed a number of methods to improve this algorithm and all these methods could be sorted into two categories, one is just based on the kNN algorithm and the other is integrating the kNN algorithm with other algorithms. In the first category, Jing et al presented a density-based method which clustered each class of sample data into several clusters and reduced the noise sample data and then combined some higher similar sample documents in each cluster into one document for reducing training data [2], Zhao et al proposed an essential vector based kNN algorithm which can keep the same accuracy with before while cutting out most of the samples [3], Bhattacharya et al put forward a new local distance function-affinity based new similarity function for kNN text classification [4], Sarma et al substituted Gaussian distribution for linear interpolation and based it on the weights to nearest neighbor [5]; while, in the second category, Ishii et al proposed a grouping method of similar words and combined it with latent semantic analysis and kNN algorithm to obtain higher accuracy in the classification [6], Zhou et al employed the hybrid model SVM-kNN algorithm which combined the advantages of SVM and kNN algorithms to improve the prediction accuracy of SVM [7], Jiang et al improved kNN algorithm on the basis of Bayesian-kNN and Citation-kNN [8].…”
Section: Introductionmentioning
confidence: 99%
“…At the worst case, if all training data are used in final algorithm decision, it will certainly spend more time. Also, more memory is required to store all training data (Chang and Liu, 2011, Jing et al, 2013, Suguna and Thanushkodi, 2010.…”
mentioning
confidence: 99%