2014
DOI: 10.48550/arxiv.1412.3730
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Inconsistency of Bayesian Inference for Misspecified Linear Models, and a Proposal for Repairing It

Abstract: We empirically show that Bayesian inference can be inconsistent under misspecification in simple linear regression problems, both in a model averaging/selection and in a Bayesian ridge regression setting. We use the standard linear model, which assumes homoskedasticity, whereas the data are heteroskedastic, and observe that the posterior puts its mass on ever more high-dimensional models as the sample size increases. To remedy the problem, we equip the likelihood in Bayes' theorem with an exponent called the l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 45 publications
0
10
0
Order By: Relevance
“…Usually, this is done for reasons completely unrelated to robustness, such as marginal likelihood approximation (Friel and Pettitt, 2008), improved MCMC mixing (Geyer, 1991), consistency in nonparametric models (Walker and Hjort, 2001;Zhang, 2006a), discounting historical data (Ibrahim and Chen, 2000), or objective Bayesian model selection (O'Hagan, 1995). However, recently, the robustness properties of power likelihoods have started to be noticed: Grünwald and van Ommen (2014) provide an in-depth study of a simulation example in which a power posterior exhibits improved robustness to misspecification, and they propose a method for choosing the power; also see Grünwald (2011Grünwald ( , 2012. Nonetheless, in all such previous research, a fixed power is used, rather than one tending to 0 as n → ∞.…”
Section: Connections With Previous Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Usually, this is done for reasons completely unrelated to robustness, such as marginal likelihood approximation (Friel and Pettitt, 2008), improved MCMC mixing (Geyer, 1991), consistency in nonparametric models (Walker and Hjort, 2001;Zhang, 2006a), discounting historical data (Ibrahim and Chen, 2000), or objective Bayesian model selection (O'Hagan, 1995). However, recently, the robustness properties of power likelihoods have started to be noticed: Grünwald and van Ommen (2014) provide an in-depth study of a simulation example in which a power posterior exhibits improved robustness to misspecification, and they propose a method for choosing the power; also see Grünwald (2011Grünwald ( , 2012. Nonetheless, in all such previous research, a fixed power is used, rather than one tending to 0 as n → ∞.…”
Section: Connections With Previous Workmentioning
confidence: 99%
“…There have been advances in robustness to misspecification, with methods such as Gibbs posteriors (Jiang and Tanner, 2008), disparity-based posteriors (Hooker and Vidyashankar, 2014), partial posteriors (Doksum and Lo, 1990), nonparametric approaches (Rodríguez and Walker, 2014), neighborhood methods (Liu and Lindsay, 2009), and learning rate adjustment (Grünwald and van Ommen, 2014); see Section 4 for a discussion of previous work. However, despite growing recognition of the issue and efforts toward a solution, existing methods tend to be either limited in scope, computationally prohibitive, or lacking a clear justification.…”
Section: Introductionmentioning
confidence: 99%
“…That is, instead of L n (β), our likelihood will be L n (β) α ; see Martin and Walker (2014). Other authors have advocated the use of a fractional likelihood, including Barron and Cover (1991), Walker and Hjort (2001), Zhang (2006), Jiang and Tanner (2008), Dalalyan and Tsybakov (2008), and Grünwald and van Ommen (2014), but these papers have different foci and none include a data-dependent (conditional) prior centering. In fact, we feel that this combination of centering and fractional likelihood regularization (see Section 2.3) is a powerful tool that can be used for a variety of high-dimensional problems.…”
Section: The Likelihood Functionmentioning
confidence: 99%
“…Our oracle property will imply that the quasi-posterior based on BMA will converge to the quasi-posterior based on the minimum risk model, asymptotically. Grünwald and van Ommen (2014) discovered suboptimal predictive performance when a homoscedastic linear model is misspecified. Their numerical experiments seem to indicate that the performance of BMA still converges to the performance of the true model eventually, albeit with a much larger sample size compared to the correctly specified case.…”
Section: Discussionmentioning
confidence: 99%
“…Our current paper only addresses the limiting distributional behavior of BMA and BMS, but not their convergence speed. As a possible future work, we may consider extending our theory in Section 3.1 to study the convergence speed in the presence of model misspecification and how the convergence depends on the temperature parameter, as discussed in Grünwald and van Ommen (2014).…”
Section: Discussionmentioning
confidence: 99%