1990
DOI: 10.1109/18.54897
|View full text |Cite
|
Sign up to set email alerts
|

Information-theoretic asymptotics of Bayes methods

Abstract: In the absence of knowledge of the true density function, Bayesian models take the joint density function for a sequence of n random variables to be an average of densities with respect to a prior. We examine the relative entropy distance D,, between the true density and the Bayesian density and show that the asymptotic distance is (d / 2 X l o g n) + c, where d is the dimension of the parameter vector. Therefore, the relative entropy rate D,,/n converges to zero at rate (l o g n) / n. The constant c, which we… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

17
357
0

Year Published

1997
1997
2014
2014

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 379 publications
(374 citation statements)
references
References 40 publications
17
357
0
Order By: Relevance
“…The already mentioned low-noise approximation to MI is constructed by employing the Cramer-Rao bound [12,[21][22][23]]. Although we demonstrate that the high-noise approximation also involves FI, we never employ the Cramer-Rao bound and the appearance of FI is due to certain asymptotic properties of the KL distance [28].…”
Section: Measures Of Informationmentioning
confidence: 99%
“…The already mentioned low-noise approximation to MI is constructed by employing the Cramer-Rao bound [12,[21][22][23]]. Although we demonstrate that the high-noise approximation also involves FI, we never employ the Cramer-Rao bound and the appearance of FI is due to certain asymptotic properties of the KL distance [28].…”
Section: Measures Of Informationmentioning
confidence: 99%
“…The asymptotics can be found analytically by expanding both bounds (7) and (9) for q → 1. For a smooth potential V both bounds give asymptotically the same logarithmic growth R Bayes m /N ≃ 1/2 ln α which can also be obtained by well known asymptotic expansions involving the Fisher information matrix [22,23,11]. On the other hand, our bounds can also be used when these standard asymptotic expansions do not apply, e.g.…”
mentioning
confidence: 68%
“…The maximum is always attained in (10) since for each at most a finite number of elements fulfill . Observe immediately the correspondence in terms of description lengths rather than probabilities Then the MDL principle is obvious: minimizes the joint description length of the model plus the data given the model 1 (see (8) and (9)). As explained before, we stick to the product notation.…”
Section: MDL Estimator and Predictionsmentioning
confidence: 99%
“…The converse holds as well. We introduce the abbreviation (8) for a semimeasure and and for the Bayes mixture . It is common to ignore the somewhat irrelevant restriction that code lengths must be an integer.…”
Section: Prerequisites and Notationmentioning
confidence: 99%