Proceedings of the Tenth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2004
DOI: 10.1145/1014052.1014115
|View full text |Cite
|
Sign up to set email alerts
|

Parallel computation of high dimensional robust correlation and covariance matrices

Abstract: The computation of covariance and correlation matrices are critical to many data mining applications and processes. Unfortunately the classical covariance and correlation matrices are very sensitive to outliers. Robust methods, such as QC and the Maronna method, have been proposed. However, existing algorithms for QC only give acceptable performance when the dimensionality of the matrix is in the hundreds; and the Maronna method is rarely used in practise because of its high computational cost. In this paper, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
1
0

Year Published

2005
2005
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 7 publications
0
1
0
Order By: Relevance
“…In this case, some value of GV cannot be compute since covariance matrix is not positive definite anymore and it becomes singular matrix. See [3], [19], [20] and [21] for further discussion regarding this problem.…”
Section: Discussionmentioning
confidence: 99%
“…In this case, some value of GV cannot be compute since covariance matrix is not positive definite anymore and it becomes singular matrix. See [3], [19], [20] and [21] for further discussion regarding this problem.…”
Section: Discussionmentioning
confidence: 99%
“…Research on unsupervised anomaly detection techniques has been actively conducted using various approaches. In the early days of the study, distribution-based anomaly detection techniques [40,52], depth-based anomaly detection techniques [53], and clustering-based anomaly detection techniques [51] have been mainly studied, but recent research trends largely utilize distanceand density-based anomaly detection techniques [54].…”
Section: Machine Learning-based Anomaly Detectionmentioning
confidence: 99%
“…A data cloud is described by a finite number of depth contours in halfspace depth. [23] Rousseeuw and Struyf (1998) introduce an algorithm to compute the halfspace depth for d = 3 with complexity O(n 2 log n). They use the Rousseeuw and Ruts (1999) procedure to determine the halfspace depth in these planes by projecting points onto planes orthogonal to the lines linking each of the points from X with z.…”
Section: Depth Function and Half Space Depthmentioning
confidence: 99%