2020
DOI: 10.1007/s11227-020-03524-3
|View full text |Cite
|
Sign up to set email alerts
|

K-DBSCAN: An improved DBSCAN algorithm for big data

Abstract: Big data storage and processing are among the most important challenges now. Among data mining algorithms, DBSCAN is a common clustering method. One of the most important drawbacks of this algorithm is its low execution speed. This study aims to accelerate the DBSCAN execution speed so that the algorithm can respond to big datasets in an acceptable period of time. To overcome the problem, an initial grouping was applied to the data in this article through the K-means++ algorithm. DBSCAN was then employed to pe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 57 publications
(13 citation statements)
references
References 30 publications
0
13
0
Order By: Relevance
“…Density-based spatial clustering of applications with noise (DBSCAN) is an unsupervised machine learning clustering algorithm [18].There are two important parameters in the DBSCAN algorithm: Eps( ) and M inP ts, the former being the neighborhood radius when defining the density and the latter being the threshold value when defining the core point [19].…”
Section: Dbscan Algorithmmentioning
confidence: 99%
“…Density-based spatial clustering of applications with noise (DBSCAN) is an unsupervised machine learning clustering algorithm [18].There are two important parameters in the DBSCAN algorithm: Eps( ) and M inP ts, the former being the neighborhood radius when defining the density and the latter being the threshold value when defining the core point [19].…”
Section: Dbscan Algorithmmentioning
confidence: 99%
“…In order to measure the clustering results of the improved method, we use Accuracy, Davies-Bouldin index (DBI), Silhouette index (Sil), Rand index (RI) [41,42], Normalized Mutual Information (NMI), Homogeneity, Completeness, and V-measure [43].…”
Section: The Error Indexmentioning
confidence: 99%
“…These high-density data are divided into different clusters. Suitable clusters are selected as references for K-value selection [ 41 ].…”
Section: Description Of the Positioning Algorithmmentioning
confidence: 99%