2016
DOI: 10.48550/arxiv.1610.04386
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Random Feature Expansions for Deep Gaussian Processes

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
3
0

Year Published

2017
2017
2019
2019

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(4 citation statements)
references
References 0 publications
1
3
0
Order By: Relevance
“…As a consequence, there is little benefit in adding additional layers after a certain point. This observation elucidates the mechanism underlying the choices of DGPs with a small number of layers for inference in numerous papers, for example in Cutajar et al (2016); Salimbeni and Deisenroth (2017); Dai et al (2015).…”
Section: Our Contributionsupporting
confidence: 63%
See 1 more Smart Citation
“…As a consequence, there is little benefit in adding additional layers after a certain point. This observation elucidates the mechanism underlying the choices of DGPs with a small number of layers for inference in numerous papers, for example in Cutajar et al (2016); Salimbeni and Deisenroth (2017); Dai et al (2015).…”
Section: Our Contributionsupporting
confidence: 63%
“…3. Following Neal (1995);Duvenaud et al (2014), recent works such as Dai et al (2015); Cutajar et al (2016) connect all layers to the input layer in order to avoid certain pathologies. The Markovian structure of the process is maintained in this case: with the above notation, the process is then defined by…”
Section: Proofmentioning
confidence: 99%
“…In this case, the gradient mixing effect will also be noticeable. We hypothesize that it could be part of the reason of the pathologies in weight-space methods; as a supporting evidence, we experimented SVGD on the random feature approximation to deep GP(Cutajar et al, 2016) on the synthetic dataset fromLouizos & Welling (2016). Using a 3-layer deep GP with Arc kernel and 5 GPs per layer, all particles collapse to a constant function, which is not the posterior mean given by HMC.…”
mentioning
confidence: 92%
“…Deep Gaussian Processes (DGP) [Damianou and Lawrence, 2013] are multilayer predictive models that are highly flexible and can accurately model uncertainty. In particular, they have been shown to perform well on a multitude of supervised regression tasks ranging from small (∼500 datapoints) to large datasets (∼500,000 datapoints) [Salimbeni and Deisenroth, 2017, Bui et al, 2016, Cutajar et al, 2016. Their main benefit over neural networks is that they are capable of capturing uncertainty in their predictions.…”
Section: Introductionmentioning
confidence: 99%