2018
DOI: 10.1109/tit.2018.2805844
|View full text |Cite
|
Sign up to set email alerts
|

Estimation of KL Divergence: Optimal Minimax Rate

Abstract: The problem of estimating the Kullback-Leibler divergence D(P Q) between two unknown distributions P and Q is studied, under the assumption that the alphabet size k of the distributions can scale to infinity. The estimation is based on m independent samples drawn from P and n independent samples drawn from Q. It is first shown that there does not exist any consistent estimator that guarantees asymptotically small worstcase quadratic risk over the set of all pairs of distributions. A restricted set that contain… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
34
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 65 publications
(34 citation statements)
references
References 25 publications
0
34
0
Order By: Relevance
“…We provide a comprehensive answer to this problem in this paper, thereby substantially generalizing the applicability of the approximation-based method and demonstrate the intricacy of functional estimation problems in high dimensions. Our recent work [16] presents the most up-to-date version of the general approximationbased method, which is applied to construct minimax rateoptimal estimators for the KL divergence (also see Bu et al [17]), χ 2 -divergence, and the squared Hellinger distance. The effective sample size enlargement phenomenon holds in all these cases as well.…”
Section: B Approximation-based Methodsmentioning
confidence: 99%
“…We provide a comprehensive answer to this problem in this paper, thereby substantially generalizing the applicability of the approximation-based method and demonstrate the intricacy of functional estimation problems in high dimensions. Our recent work [16] presents the most up-to-date version of the general approximationbased method, which is applied to construct minimax rateoptimal estimators for the KL divergence (also see Bu et al [17]), χ 2 -divergence, and the squared Hellinger distance. The effective sample size enlargement phenomenon holds in all these cases as well.…”
Section: B Approximation-based Methodsmentioning
confidence: 99%
“…Classically, there does not exist any consistent estimator that guarantees asymptotically small error over the set of all pairs of distributions [14,27]. These two papers then consider pairs of distributions with bounded probability ratios specified by a function f : N → R + , namely all pairs of distributions in the set as follows:…”
Section: Application: Kl Divergence Estimationmentioning
confidence: 99%
“…Denote the number of samples from p and q to be M p and M q , respectively. References [14,27] shows that classically, D KL (p q) can be approximated within constant additive error with high success probability if and only if M p = Ω( n log n ) and M q = Ω( nf (n) log n ). Quantumly, we are given unitary oraclesÔ p andÔ q defined by (1.7).…”
Section: Application: Kl Divergence Estimationmentioning
confidence: 99%
“…Although we focus on Bayesian machinery, one can estimate D-probabilities using any method that estimates KL(f 0 , f j ). Substantial work has focused on estimating the Kullback-Leibler divergence between two unknown densities based on samples from these densities (Leonenko et al, 2008;Pérez-Cruz, 2008;Bu et al, 2018). Our setting is somewhat different, but the local likelihood methods of Lee and Park (2006) and the Bayesian approach of Viele (2007) can potentially be used, among others.…”
Section: Estimation Of Model Weights: D-probabilitiesmentioning
confidence: 99%