2015 IEEE International Symposium on Information Theory (ISIT) 2015
DOI: 10.1109/isit.2015.7282679
|View full text |Cite
|
Sign up to set email alerts
|

Does dirichlet prior smoothing solve the Shannon entropy estimation problem?

Abstract: Abstract-We consider the problem of estimating functionals of discrete distributions, and focus on tight (up to universal multiplicative constants for each specific functional) nonasymptotic analysis of the worst case squared error risk of widely used estimators. We apply concentration inequalities to analyze the random fluctuation of these estimators around their expectations, and the theory of approximation using positive linear operators to analyze the deviation of their expectations from the true functiona… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
16
0

Year Published

2015
2015
2019
2019

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 12 publications
(16 citation statements)
references
References 48 publications
0
16
0
Order By: Relevance
“…Wolpert and Wolf [56] gives an explicit expression of this estimator: H^Bayes=truei=1Sn0.2emp^i+an+Sa·false(ψ0false(n+Sa+1false)ψ0false(n0.2emp^i+a+1false)false).It has been shown in [58] that this approach cannot achieve the minimax rates. We feed the algorithm with true S and set a=n/S in our experiments.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…Wolpert and Wolf [56] gives an explicit expression of this estimator: H^Bayes=truei=1Sn0.2emp^i+an+Sa·false(ψ0false(n+Sa+1false)ψ0false(n0.2emp^i+a+1false)false).It has been shown in [58] that this approach cannot achieve the minimax rates. We feed the algorithm with true S and set a=n/S in our experiments.…”
Section: Methodsmentioning
confidence: 99%
“…Bearing this in mind, Figure 5 suggests that the following three estimators out of 12, namely our estimator, the estimator by Valiant and Valiant [25] and the best upper bound (BUB) estimator, achieve the minimax rate. Some other estimators have already been shown not to achieve the optimal sample complexity, e.g., the Miller-Madow bias-corrected MLE [29], [52], the jackknifed estimator [52], and the Dirichlet-smoothed plug-in estimator as well as the Bayes estimator under Dirichlet prior [58]. We remark that the shrinkage estimator only improves the MLE when the distribution is near-uniform, and this estimator performs poorly under the Zipf distribution (verified by our experiments), thus it attains neither the optimal minimum sample complexity nor the minimax rate.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Besides, the estimator based on the best polynomial approximation also has the same performance [123]. Moreover, an inferior estimator is constructed by use of Dirichlet prior smoothing, which is similar to MLE but not as good as the above two [124]. In addition, an ensemble of plug-in estimators with weights is proposed to protect the results of estimation from decaying with the increase of sample dimension [125].…”
Section: A Efficient Estimation Of Information Measuresmentioning
confidence: 99%
“…- [24] still inconsistent if k is larger than n. The estimators based on Bayesian approaches in [16]- [18] are also inconsistent for k n [19]. Paninski [20] firstly revealed existence of a consistent estimator even if the alphabet size k is larger than linear order of the sample size n. However, they did not provide a concrete form of the consistent estimator.…”
Section: A Related Workmentioning
confidence: 99%