2017
DOI: 10.1007/s11222-016-9721-7
|View full text |Cite
|
Sign up to set email alerts
|

Large-scale kernel methods for independence testing

Abstract: Representations of probability measures in reproducing kernel Hilbert spaces provide a flexible framework for fully nonparametric hypothesis tests of independence, which can capture any type of departure from independence, including nonlinear associations and multivariate interactions. However, these approaches come with an at least quadratic computational cost in the number of observations, which can be prohibitive in many applications. Arguably, it is exactly in such large-scale datasets that capturing any t… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
79
0

Year Published

2017
2017
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 76 publications
(80 citation statements)
references
References 27 publications
1
79
0
Order By: Relevance
“…6. It might be possible to reduce computational complexity by using linear time approximation methods as described by Zhang et al (2017) for the pairwise HSIC. (We thank one of the referees for pointing this out.…”
Section: Run Time Analysismentioning
confidence: 99%
“…6. It might be possible to reduce computational complexity by using linear time approximation methods as described by Zhang et al (2017) for the pairwise HSIC. (We thank one of the referees for pointing this out.…”
Section: Run Time Analysismentioning
confidence: 99%
“…Block HSIC Lasso employs the block HSIC estimator (Zhang et al, 2018) instead of the V-statistics estimator of Equation (2). More specifically, to compute the block HSIC, we first partition the training…”
Section: Block Hsic Lassomentioning
confidence: 99%
“…In this paper, we propose block HSIC Lasso: a simple yet effective nonlinear feature selection algorithm based on HSIC Lasso. The key idea is to use the recently proposed block HSIC estimator (Zhang et al, 2018) to estimate the HSIC terms. By splitting the data in blocks of size B ⌧ n, the memory complexity of HSIC Lasso goes from O(dn 2 ) down to O(dnB).…”
Section: Introductionmentioning
confidence: 99%
“…Non-parametric tests have no restrictions in the sample inference process. The test used here employs a statistical permutation (Collingridge, 2013;Konietschke and Pauly, 2014;Koopman et al, 2015;Zhang et al, 2017;Pauly et al, 2018;Derrick et al 2018).…”
Section: Validation Of the Simulationsmentioning
confidence: 99%