2020
DOI: 10.1109/access.2020.2974764
|View full text |Cite
|
Sign up to set email alerts
|

A Distributed Storage and Computation k-Nearest Neighbor Algorithm Based Cloud-Edge Computing for Cyber-Physical-Social Systems

Abstract: The k-nearest neighbor (kNN) algorithm is a classic supervised machine learning algorithm. It is widely used in cyber-physical-social systems (CPSS) to analyze and mine data. However, in practical CPSS applications, the standard linear kNN algorithm struggles to efficiently process massive data sets. This paper proposes a distributed storage and computation k-nearest neighbor (D-kNN) algorithm. The D-kNN algorithm has the following advantages: First, the concept of k-nearest neighbor boundaries is proposed and… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
23
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 51 publications
(24 citation statements)
references
References 34 publications
0
23
0
1
Order By: Relevance
“…In both tasks, the input features consists of K closest training examples or the dataset in the feature space, while the algorithm relies on labeled data for the learning process to produce appropriate outputs for unlabeled input features. The idea behind the KNN algorithm is that if a sample has k most similar neighbors in the feature space, m13ost of the samples belong to a certain category, then the sample also belongs to this category [ 4 , 30 , 31 ]. The voting method is generally used in the classification task, that is, the category label that appears frequently in the k sample is selected as the prediction result, while in the case of the regression task, the average method is used where the real value output labels of the k sample are used as the prediction result [ 5 , 31 ].…”
Section: Methodsmentioning
confidence: 99%
“…In both tasks, the input features consists of K closest training examples or the dataset in the feature space, while the algorithm relies on labeled data for the learning process to produce appropriate outputs for unlabeled input features. The idea behind the KNN algorithm is that if a sample has k most similar neighbors in the feature space, m13ost of the samples belong to a certain category, then the sample also belongs to this category [ 4 , 30 , 31 ]. The voting method is generally used in the classification task, that is, the category label that appears frequently in the k sample is selected as the prediction result, while in the case of the regression task, the average method is used where the real value output labels of the k sample are used as the prediction result [ 5 , 31 ].…”
Section: Methodsmentioning
confidence: 99%
“…Ada juga yang berpendapat bahwa algoritma KNN adalah algoritma pembelajaran yang banyak digunakan dalam sistem cyber-fisiksosial (CPSS) untuk menganalisis dan menambang data (main data). (Zhang, Chen, Liu, & Xi, 2020).…”
Section: Pendahuluanunclassified
“…Data-driven proactive management will be essentially supported by ML/AI techniques. These techniques, once deployed in future networked systems, should offer a new range of networking-based services such as smart routing in networks with a cross-layer design [59], task offloading and resource allocation [60], optimized operation of next-generation mobile networks [61], distributed storage and computation at the network edge [62], and accurate localization estimation of mobile robots [63]. In parallel with this expected network evolution, novel challenges such as privacy, e.g.…”
Section: Data-triggered Management Mechanismmentioning
confidence: 99%