2019
DOI: 10.48550/arxiv.1912.00894
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On the geometry of Stein variational gradient descent

Abstract: Bayesian inference problems require sampling or approximating high-dimensional probability distributions. The focus of this paper is on the recently introduced Stein variational gradient descent methodology, a class of algorithms that rely on iterated steepest descent steps with respect to a reproducing kernel Hilbert space norm. This construction leads to interacting particle systems, the mean-field limit of which is a gradient flow on the space of probability distributions equipped with a certain geometrical… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
60
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 25 publications
(60 citation statements)
references
References 78 publications
0
60
0
Order By: Relevance
“…Constructing numerical approximations for ( 6)-( 7) beyond the constant-in-space approximation is a topic of ongoing research. We mention in particular the approach developed in [97] and analysed in [85] based on diffusion maps as well as the method from [79] based on the Stein geometry [41,81].…”
Section: From the Filtering Problem To The Mckean-vlasov Equationmentioning
confidence: 99%
“…Constructing numerical approximations for ( 6)-( 7) beyond the constant-in-space approximation is a topic of ongoing research. We mention in particular the approach developed in [97] and analysed in [85] based on diffusion maps as well as the method from [79] based on the Stein geometry [41,81].…”
Section: From the Filtering Problem To The Mckean-vlasov Equationmentioning
confidence: 99%
“…The third option we consider is the non-parametric updates, which we derive approximating the energy gradient in RKHS H with kernel k. To minimize the KL-divergence we first take the Frechet derivative w.r.t. the energy E along some direction h and use the fact that H is actually dense in L 2 q (Duncan et al, 2019). Then, using the reproducing property of k, we can formulate the directional derivative as an action of a linear operator:…”
Section: Algorithm 1 Methods α βmentioning
confidence: 99%
“…Stein Variational Gradient Descent (SVGD) Liu and Wang [2016], Liu [2017] is an alternative to the Langevin algorithm and has been applied in several contexts in machine learning, including Reinforcement Learning , sequential decision making Zhang et al [2018, Generative Adversarial Networks Tao et al [2019], Variational Auto Encoders Pu et al [2017], and Federated Learning Kassab and Simeone [2020]. However, the theoretical understanding of SVGD is limited compared to that of Langevin algorithm Lu et al [2019], Duncan et al [2019], Liu [2017], Chewi et al [2020], Nüsken and Renger [2021]. In particular, the first complexity result of SVGD, due to Korba et al [2020, Corollary 6], appeared only recently, and relies on an assumption on the trajectory of the algorithm, which cannot be checked prior to running the algorithm.…”
Section: Stein Variational Gradient Descent (Svgd)mentioning
confidence: 99%