2020
DOI: 10.48550/arxiv.2002.09547
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Stochastic Normalizing Flows

Liam Hodgkinson,
Chris van der Heide,
Fred Roosta
et al.

Abstract: We introduce stochastic normalizing flows, an extension of continuous normalizing flows for maximum likelihood estimation and variational inference (VI) using stochastic differential equations (SDEs). Using the theory of rough paths, the underlying Brownian motion is treated as a latent variable and approximated, enabling efficient training of neural SDEs as random neural ordinary differential equations. These SDEs can be used for constructing efficient Markov chains to sample from the underlying distribution … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
11
0

Year Published

2020
2020
2021
2021

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 7 publications
(11 citation statements)
references
References 23 publications
0
11
0
Order By: Relevance
“…These continuous-depth models give rise to vector field representations and a set of functions which were not possible to generate before. These capabilities enabled flexible density estimation (Dupont et al, 2019, Grathwohl et al, 2018, Hodgkinson et al, 2020, Mathieu and Nickel, 2020, Yang et al, 2019, as well as performant modeling of sequential and irregularly-sampled data (Erichson et al, 2021, Gholami et al, 2019, Lechner and Hasani, 2020, Rubanova et al, 2019. In this paper, we showed how to relax the need for an ODE-solver to realize an expressive continuous-time neural network model for challenging time-series problems.…”
Section: Related Workmentioning
confidence: 93%
“…These continuous-depth models give rise to vector field representations and a set of functions which were not possible to generate before. These capabilities enabled flexible density estimation (Dupont et al, 2019, Grathwohl et al, 2018, Hodgkinson et al, 2020, Mathieu and Nickel, 2020, Yang et al, 2019, as well as performant modeling of sequential and irregularly-sampled data (Erichson et al, 2021, Gholami et al, 2019, Lechner and Hasani, 2020, Rubanova et al, 2019. In this paper, we showed how to relax the need for an ODE-solver to realize an expressive continuous-time neural network model for challenging time-series problems.…”
Section: Related Workmentioning
confidence: 93%
“…Exploiting the connections between ODE-Nets and the theory of dynamical systems and control is an active area of research [50,53,68], that has also motivated the development of more memory efficient training strategies [13,19,20,54,72] for ODE-Nets. Other research fronts include normalizing flows [21,36,65], and stochastic differential equations [23,32,34,42,45].…”
Section: Related Workmentioning
confidence: 99%
“…We consider a model that has 3 ODE-blocks. The units within the three blocks have an increasing number of channels, [16,32,64].…”
Section: Compressing Deep Ode-nets For Image Classification Tasksmentioning
confidence: 99%
“…A related work may be found in [64], where equations for time-varying cumulative distribution functions could directly be learned given access to a fine-scaled simulator. Another relevant work in the literature was presented in [21] where SDEs are utilized to parameterize the transformation of a sample from a latent (generally a time-invariant multivariate isotropic Gaussian) to a time-varying target density. It is important to clarify here that concatenated transformations between supports can also be devised as flow maps governed by ordinary differential equations (ODEs) or SDEs (like in [21]).…”
Section: Introductionmentioning
confidence: 99%
“…Another relevant work in the literature was presented in [21] where SDEs are utilized to parameterize the transformation of a sample from a latent (generally a time-invariant multivariate isotropic Gaussian) to a time-varying target density. It is important to clarify here that concatenated transformations between supports can also be devised as flow maps governed by ordinary differential equations (ODEs) or SDEs (like in [21]). Note that the solutions of these equations are obtained through numerical discretization in fictitious time and must not be confused with the physical time evolution of the target densities themselves; evolution in this fictitious time is then learned and deployed to construct each reference-to-target transformation.…”
Section: Introductionmentioning
confidence: 99%