2018
DOI: 10.48550/arxiv.1806.05490
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Inference in Deep Gaussian Processes using Stochastic Gradient Hamiltonian Monte Carlo

Marton Havasi,
José Miguel Hernández-Lobato,
Juan José Murillo-Fuentes

Abstract: Deep Gaussian Processes (DGPs) are hierarchical generalizations of Gaussian Processes that combine well calibrated uncertainty estimates with the high flexibility of multilayer models. One of the biggest challenges with these models is that exact inference is intractable. The current state-of-the-art inference method, Variational Inference (VI), employs a Gaussian approximation to the posterior distribution. This can be a potentially poor unimodal approximation of the generally multimodal posterior. In this wo… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(5 citation statements)
references
References 6 publications
0
5
0
Order By: Relevance
“…Moreover, our experiments involving more complicated architectures, such as ResNet or DenseNet for standard multi-class classification, have not managed to surpass in accuracy a far less complex model with only 3 hidden layers. A plausible reason behind this under-fitting resides in the factorized approximate posterior formulation, which was shown to negatively affect predictive performance compared to MCMC inference schemes (Havasi et al, 2018). We posit that using alternative inference frameworks (Ustyuzhaninov et al, 2019) whereby we impose correlations between layers might alleviate this issue.…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, our experiments involving more complicated architectures, such as ResNet or DenseNet for standard multi-class classification, have not managed to surpass in accuracy a far less complex model with only 3 hidden layers. A plausible reason behind this under-fitting resides in the factorized approximate posterior formulation, which was shown to negatively affect predictive performance compared to MCMC inference schemes (Havasi et al, 2018). We posit that using alternative inference frameworks (Ustyuzhaninov et al, 2019) whereby we impose correlations between layers might alleviate this issue.…”
Section: Discussionmentioning
confidence: 99%
“…Moreover, our experiments involving more complicated architectures, such as ResNet or DenseNet for standard multi-class classification, have not managed to surpass in accuracy a far less complex model with only 3 hidden layers. A plausible reason behind this under-fitting resides in the factorized approximate posterior formulation, which was shown to negatively affect predictive performance compared to MCMC inference schemes (Havasi et al, 2018). We posit that using alternative inference frameworks (Ustyuzhaninov et al, 2019) whereby we impose correlations between layers might alleviate this issue.…”
Section: Discussionmentioning
confidence: 99%
“…e Stochastic Gradient Hamiltonian Monte Carlo approach (Ma et al, 2015) has proven e cient in deep GPs (Havasi et al, 2018) and in GANs (Saatci and Wilson, 2017). Another avenue for improvement lies in kernel interpolation techniques (Wilson and Nickisch, 2015;Evans and Nair, 2018) which would make inference and prediction faster.…”
Section: Discussionmentioning
confidence: 99%