2016 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and I 2016
DOI: 10.1109/ithings-greencom-cpscom-smartdata.2016.177
|View full text |Cite
|
Sign up to set email alerts
|

From Big Data to Smart Data with the K-Nearest Neighbours Algorithm

Abstract: Abstract-The k-nearest neighbours algorithm is one of the most widely used data mining models because of its simplicity and accurate results. However, when it comes to deal with big datasets, with potentially noisy and missing information, this technique becomes ineffective and inefficient. Due to its drawbacks to tackle large amounts of imperfect data, plenty of research has aimed at improving this algorithm by means of data preprocessing techniques. These weaknesses have turned out as strengths and the k-nea… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
5
2
1

Relationship

2
6

Authors

Journals

citations
Cited by 15 publications
(6 citation statements)
references
References 23 publications
0
3
0
3
Order By: Relevance
“…Although most of these data preprocessing techniques were motivated by k‐NN drawbacks, it turns out that the resulting “smart” dataset provided by the above approaches can also be of use in many other learning algorithms Cano, Herrera, and Lozano (); Luengo et al (). This work reviews the current specialized literature that revolves around the idea of the k‐NN to come up with Smart Data, greatly extending our preliminary contribution in Triguero, Maillo, Luengo, García, and Herrera () around this topic. First, we will deepen into the concepts of big and Smart Data and how to extract value from Big Data with existing technologies and Big Data preprocessing techniques (Section 2).…”
Section: Introductionmentioning
confidence: 87%
“…Although most of these data preprocessing techniques were motivated by k‐NN drawbacks, it turns out that the resulting “smart” dataset provided by the above approaches can also be of use in many other learning algorithms Cano, Herrera, and Lozano (); Luengo et al (). This work reviews the current specialized literature that revolves around the idea of the k‐NN to come up with Smart Data, greatly extending our preliminary contribution in Triguero, Maillo, Luengo, García, and Herrera () around this topic. First, we will deepen into the concepts of big and Smart Data and how to extract value from Big Data with existing technologies and Big Data preprocessing techniques (Section 2).…”
Section: Introductionmentioning
confidence: 87%
“…This method is widely applied for data classification and pattern recognition [26]. KNN is a high cost technique for have to find all the distances between data [27] and K the centroid to find out the closest distance that represents the similarity of the data to a certain group.…”
Section: ) Association Rulementioning
confidence: 99%
“…Omenjeni proces predstavlja eno izmed pomembnih faz podatkovnega rudarjenja, namenjeno čiščenju in korigiranju izvornih podatkov z namenom učinkovitejšega apliciranja algoritmov strojnega učenja (npr. evolucijskih algoritmov, globokega učenja, regresijske analize) [23]. Med opravila predpriprave podatkov štejemo čiščenje (angl.…”
Section: Strojno Učenjeunclassified