2020
DOI: 10.1007/978-3-030-47426-3_27
|View full text |Cite
|
Sign up to set email alerts
|

Decentralized and Adaptive K-Means Clustering for Non-IID Data Using HyperLogLog Counters

Abstract: The data shared over the Internet tends to originate from ubiquitous and autonomous sources such as mobile phones, fitness trackers, and IoT devices. Centralized and federated machine learning solutions represent the predominant way of providing smart services for users. However, moving data to central location for analysis causes not only many privacy concerns, but also communication overhead. Therefore, in certain situations machine learning models need to be trained in a collaborative and decentralized mann… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 10 publications
(5 citation statements)
references
References 16 publications
0
5
0
Order By: Relevance
“…An alternative private distributed clustering approach is to select a subset of local points (representatives) and apply clustering over them. Soliman et al [21] proposed running the K-Means algorithm locally on each client and using HyperLogLog counters to share the centroids and the approximate number of observations per centroid in a decentralized fashion with the other clients. Then a weighted averaging over all the centroids is done to find the global centroids.…”
Section: Privacy Preserving Distributed K-meansmentioning
confidence: 99%
“…An alternative private distributed clustering approach is to select a subset of local points (representatives) and apply clustering over them. Soliman et al [21] proposed running the K-Means algorithm locally on each client and using HyperLogLog counters to share the centroids and the approximate number of observations per centroid in a decentralized fashion with the other clients. Then a weighted averaging over all the centroids is done to find the global centroids.…”
Section: Privacy Preserving Distributed K-meansmentioning
confidence: 99%
“…• To aggregate the information learned from each of the clients into a global model using classical federated aggregation operators such as: FedAvg, weighted FedAvg [24] and an aggregation operator for the adaptation of the k-means algorithm to a federated setting [25].…”
Section: Software Functionalitiesmentioning
confidence: 99%
“…Although there are many works concentrating on the distributed learning with clustering as methods to compute local model [17,18,19,20,21], few works have considered the performance aspects of ML models in an inherently distributed network, especially when there are sub-groups of agents receiving data from different phenomena. At some timestamps T i , some agents may have learned their provisional global models or received model from their neighbours.…”
Section: Introductionmentioning
confidence: 99%