2017
DOI: 10.1145/3068335
|View full text |Cite
|
Sign up to set email alerts
|

DBSCAN Revisited, Revisited

Abstract: At SIGMOD 2015, an article was presented with the title “DBSCAN Revisited: Mis-Claim, Un-Fixability, and Approximation” that won the conference’s best paper award. In this technical correspondence, we want to point out some inaccuracies in the way DBSCAN was represented, and why the criticism should have been directed at the assumption about the performance of spatial index structures such as R-trees and not at an algorithm that can use such indexes. We will also discuss the relationship of DBSCAN performance … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
534
0
9

Year Published

2017
2017
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 1,810 publications
(545 citation statements)
references
References 34 publications
2
534
0
9
Order By: Relevance
“…2 Aii), additional QC steps were required. The traces were first 159 clustered using DBSCAN [16,17], and each cluster median was compared against white 160 noise of the same mean and standard deviation (KS-test), and rejected as artifact if it 161 was not significantly p < 0.05 different from Gaussian white noise of matched mean and 162 variance, e.g. Fig.…”
Section: /32mentioning
confidence: 99%
“…2 Aii), additional QC steps were required. The traces were first 159 clustered using DBSCAN [16,17], and each cluster median was compared against white 160 noise of the same mean and standard deviation (KS-test), and rejected as artifact if it 161 was not significantly p < 0.05 different from Gaussian white noise of matched mean and 162 variance, e.g. Fig.…”
Section: /32mentioning
confidence: 99%
“…In cases where such noisy data exists, it is necessary to use more robust methods such as Gaussian mixture models (GMMs) 48 or hierarchical density-based spatial clustering of applications with noise (DBSCAN). 49 Some machine learning models do not perform well with noisy data compared with others such as ensemble methods. 50 Ensemble methods aggregate prediction results through multiple models (eg, decision trees) and then average the differences among these multiple models.…”
Section: Missing Data Tolerancementioning
confidence: 99%
“…The implementation of the final clustering is omitted at this point and widely described in literature [36]. With the original algorithm implemented, adaptions only have to be made in the core point definition and neighborhood metrics.…”
Section: Modified Core Point Definition For Dbstacmentioning
confidence: 99%