2017 IEEE International Symposium on Information Theory (ISIT) 2017
DOI: 10.1109/isit.2017.8007085
|View full text |Cite
|
Sign up to set email alerts
|

Minimax risk for missing mass estimation

Abstract: Abstract-The problem of estimating the missing mass or total probability of unseen elements in a sequence of n random samples is considered under the squared error loss function. The worstcase risk of the popular Good-Turing estimator is shown to be between 0.6080/n and 0.6179/n. The minimax risk is shown to be lower bounded by 0.25/n. This appears to be the first such published result on minimax risk for estimation of missing mass, which has several practical and theoretical applications.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
4

Relationship

1
8

Authors

Journals

citations
Cited by 21 publications
(14 citation statements)
references
References 19 publications
0
14
0
Order By: Relevance
“…Remarks: As we show in Section VII-A, Var(M 0 ) ≤ γ/n, and there exist uniform distributions that nearly achieve this variance upper bound [19]. Now, if Z ∼ sub-Gaussian(v), it is known that v ≥ var(Z) [24].…”
Section: B Concentration Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…Remarks: As we show in Section VII-A, Var(M 0 ) ≤ γ/n, and there exist uniform distributions that nearly achieve this variance upper bound [19]. Now, if Z ∼ sub-Gaussian(v), it is known that v ≥ var(Z) [24].…”
Section: B Concentration Resultsmentioning
confidence: 99%
“…Paraphrasing the first part of the theorem, the minimax risk of the order-α missing mass for a positive integer α falls as 1/n 2α−1 . In a proof presented in Section IV, we show the lower bound in (19) using Dirichlet priors and the upper bound using a generalized Good-Turing estimator for M 0,α (X n , P ), denoted M GT 0,α (X n ), defined as…”
Section: Summary Of Resultsmentioning
confidence: 99%
“…As mentioned in the introduction, a good part of the work on learning the missing mass focused on additive error [ 2 , 3 , 4 ]. Recently, minimax lower bounds were given for the additive error in [ 11 ] and [ 12 ]. Note however that relative error bounds cannot be deduced from these (nor any other way, given the impossibility established here.)…”
Section: Discussionmentioning
confidence: 99%
“…We relate the Good-Turing estimator for feature allocation models with the classical Good-Turing estimator for species sampling, by relying on the limiting Poisson approximation of Binomial random variables. In particular, Theorem 2.1 states that, in the feature allocation models, the Good-Turing estimator achieves a risk of order R n ≍ W n , while it is known from Rajaraman et al (2017) that the riskR n of the Good Turing estimator in the species sampling case asymptotically behaves as R n ≍ 1/n. In order to compare the two models, we will consider the limiting scenario when W → 0.…”
Section: Connection To the Good Turing Estimator For Species Samplingmentioning
confidence: 99%