2016
DOI: 10.1080/00401706.2015.1054435
|View full text |Cite|
|
Sign up to set email alerts
|

Fast Computing for Distance Covariance

Abstract: Distance covariance and distance correlation have been widely adopted in measuring dependence of a pair of random variables or random vectors. If the computation of distance covariance and distance correlation is implemented directly accordingly to its definition then its computational complexity is O(n 2 ) which is a disadvantage compared to other faster methods. In this paper we show that the computation of distance covariance and distance correlation of real valued random variables can be implemented by an … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
97
0

Year Published

2017
2017
2023
2023

Publication Types

Select...
7
1

Relationship

2
6

Authors

Journals

citations
Cited by 112 publications
(97 citation statements)
references
References 13 publications
0
97
0
Order By: Relevance
“…The dCor also has quadratic complexity though it is hard to see from our figure. (Though the R implementation of dCor, which we adopt here, has quadratic complexity in sample size, we note that there is a recent work (Huo and Székely, 2016) proposes a new algorithm for computing dCor that has computational complexity O(n log n) in sample size.) To see how the computation under FES scale with the marginal maximum resolution k 1 and k 2 , we repeat the analysis for k 1 = k 2 = 4, 5, .…”
Section: Computational Scalabilitymentioning
confidence: 99%
“…The dCor also has quadratic complexity though it is hard to see from our figure. (Though the R implementation of dCor, which we adopt here, has quadratic complexity in sample size, we note that there is a recent work (Huo and Székely, 2016) proposes a new algorithm for computing dCor that has computational complexity O(n log n) in sample size.) To see how the computation under FES scale with the marginal maximum resolution k 1 and k 2 , we repeat the analysis for k 1 = k 2 = 4, 5, .…”
Section: Computational Scalabilitymentioning
confidence: 99%
“…However, sparse data is the relevant context for the PDC periodogram, since in large datasets periodicities are usually adequately detectable by the GLS or other conventional techniques. Nevertheless, one approach to try and improve the complexity of the computation is suggested in Huo & Székely (2016). Huo & Székely use an unbiased version of distance correlation first suggested by Székely & Rizzo (2014), and apply to it an AVL-tree computational approach (Adelson-Velskii & Landis 1962), to obtain an O(N log N) algorithm to calculate the distance correlation.…”
Section: Discussionmentioning
confidence: 99%
“…Here, X ′ (respectively Y ′ ) denotes an independent copy of X (respectively Y ). An unbiased estimator of squared distance covariance proposed by Székely and Rizzo () is given by V^U(X,Y)=1n(n3)ijÃijB˜ij for n > 3, where the so‐called scriptU‐centred matrices, Ãij, have the additional property that Efalse[Ãijfalse]=0 for all i , j and are defined by Ãij=aij1n2truel=1nail1n2truek=1nakj+1false(n1false)false(n2false)truek,l=1nakl,ij;0,i=j. For p = q = 1, Huo and Székely () have shown that trueV^Ufalse(X,Yfalse) is a U‐statistic, which is degenerate in the case where X and Y are independent. For further details concerning the properties of trueV^Ufalse(boldX,boldYfalse), we refer to the paper by Huang and Huo ().…”
Section: Estimation Testing and Further Propertiesmentioning
confidence: 99%
“…For further details concerning the properties of trueV^Ufalse(boldX,boldYfalse), we refer to the paper by Huang and Huo (). By using the U‐statistic representation of trueV^Ufalse(X,Yfalse), Huo and Székely () show that it can be computed by an Ofalse(nnormallognfalse) algorithm. This algorithm considerably speeds up the calculation of the distance correlation coefficient for large sample sizes.…”
Section: Estimation Testing and Further Propertiesmentioning
confidence: 99%