2016
DOI: 10.1007/s00180-015-0633-3
|View full text |Cite
|
Sign up to set email alerts
|

Two-sample homogeneity tests based on divergence measures

Abstract: The concept of f -divergences introduced by [2] provides a rich set of distance like measures between pairs of distributions. Divergences do not focus on certain moments of random variables, but rather consider discrepancies between the corresponding probability density functions. Thus, two-sample tests based on these measures can detect arbitrary alternatives when testing the equality of the distributions. We treat the problem of divergence estimation as well as the subsequent testing for the homogeneity of t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 13 publications
(12 citation statements)
references
References 23 publications
0
12
0
Order By: Relevance
“…Thus, the two-sample test statistic between Ŝ and S measures the fidelity of the samples Ŝ produced by the generative model. The use of two-sample tests to evaluate the sample quality of generative models include the pioneering work of Box (1980), the use of Maximum Mean Discrepancy (MMD) criterion (Bengio et al, 2013;Dziugaite et al, 2015;Lloyd & Ghahramani, 2015;Bounliphone et al, 2015;Sutherland et al, 2016), and the connections to density-ratio estimation (Kanamori et al, 2010;Wornowizki & Fried, 2016;Menon & Ong, 2016;Mohamed & Lakshminarayanan, 2016).…”
Section: Two-sample Testingmentioning
confidence: 99%
“…Thus, the two-sample test statistic between Ŝ and S measures the fidelity of the samples Ŝ produced by the generative model. The use of two-sample tests to evaluate the sample quality of generative models include the pioneering work of Box (1980), the use of Maximum Mean Discrepancy (MMD) criterion (Bengio et al, 2013;Dziugaite et al, 2015;Lloyd & Ghahramani, 2015;Bounliphone et al, 2015;Sutherland et al, 2016), and the connections to density-ratio estimation (Kanamori et al, 2010;Wornowizki & Fried, 2016;Menon & Ong, 2016;Mohamed & Lakshminarayanan, 2016).…”
Section: Two-sample Testingmentioning
confidence: 99%
“…In information theory, importance weight estimation (where it is sometimes known as Radon-Nikodym derivative estimation), is applied to estimate various divergences such as the KL divergence and Renyi divergences between the distributions P and R [NWJ07, NWJ10, YSK + 13, WKV05, WKV09]. These divergence measures themselves find several applications, such as two-sample tests for distinguishing between two distributions [WF16] and for independence testing [SSSK08].…”
Section: Introductionmentioning
confidence: 99%
“…Training classifier provides an estimator of density ratio, as has been pointed out in [27] and in the formulation of learning generative models [28]. While distribution divergence estimation has been studied and used for two-sample problems [18,19,37,40], the use of neural network as a divergence estimator for two-sample testing was less investigated. In terms of theoretical guarantee of test power, the analysis in [26] assumes a non-zero population test statistic under H 1 which is not specified, along with other approximations.…”
Section: Related Workmentioning
confidence: 99%
“…Many existing tests are based on certain estimators of a distance or divergence between p and q. Important examples include Maximum Mean Discrepancy (MMD), especially kernel-based MMD [1,13] and distance of Reproducing Kernel Hilbert Space Mean Embedding [9,16], divergence based methods which may involve non-parametric estimation of density difference or density ratio [19,36,37,40]. While these methods have been intensively studied and theoretically well-understood, the application is often restricted to data of small dimensionality and/or small sample size, or certain specific classes of densities p and q, due to model and computational limitations.…”
Section: Introductionmentioning
confidence: 99%