2019
DOI: 10.48550/arxiv.1902.09754
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Function Space Particle Optimization for Bayesian Neural Networks

Abstract: While Bayesian neural networks (BNNs) have drawn increasing attention, their posterior inference remains challenging, due to the high-dimensional and overparameterized nature. Recently, several highly flexible and scalable variational inference procedures based on the idea of particle optimization have been proposed. These methods directly optimize a set of particles to approximate the target posterior. However, their application to BNNs often yields sub-optimal performance, as they have a particular failure m… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(10 citation statements)
references
References 18 publications
0
10
0
Order By: Relevance
“…Kendall & Gal (2017) used MC dropout for model uncertainty and combined it with the idea by Nix & Weigend (1994) of directly modeling mean and data noise as network outputs. Wang et al (2019) and Wen et al (2020) refined ensemble methods for NNs by further promoting their diversity on the function space and by reducing their computational cost, respectively. For classification, Malinin & Gales (2018) introduced prior networks, which explicitly model in-sample and out-of-sample uncertainty, where the latter is realized by minimizing the reverse KL-distance to a selected flat pointwise defined prior.…”
Section: Overview Of Our Contributionmentioning
confidence: 99%
“…Kendall & Gal (2017) used MC dropout for model uncertainty and combined it with the idea by Nix & Weigend (1994) of directly modeling mean and data noise as network outputs. Wang et al (2019) and Wen et al (2020) refined ensemble methods for NNs by further promoting their diversity on the function space and by reducing their computational cost, respectively. For classification, Malinin & Gales (2018) introduced prior networks, which explicitly model in-sample and out-of-sample uncertainty, where the latter is realized by minimizing the reverse KL-distance to a selected flat pointwise defined prior.…”
Section: Overview Of Our Contributionmentioning
confidence: 99%
“…We study different kernel functions in the weight and function space, as well as deterministic and stochastic update rules. This includes some existing approaches as special cases, such as standard deep ensembles [33], BNN weight-space SVGD [26], and BNN function-space SVGD [55], but it also includes several novel approaches. We will lay out the motivation and theoretical properties for each approach and later proceed to empirically evaluating their respective performance.…”
Section: Stein Variational Neural Network Ensemblesmentioning
confidence: 99%
“…theoretically and empirically. We also include two new approaches in our comparison and show that our hybrid h-SVGD method, that acts both in the weight and function space, and our fw-SVGD method, that fixes an issue with an existing functional SVGD approach [55], lead to more diverse ensembles and improved uncertainty estimation and out-of-distribution detection, as well as approaching the gold-standard Hamiltonian Monte Carlo posterior more closely.…”
Section: Introductionmentioning
confidence: 99%
“…In the context of machine learning a number of recent works have proposed gradient flow formulations of methods for sampling and variational inference, see for example [4,40,44,50,86,88].…”
Section: Previous Workmentioning
confidence: 99%