2019
DOI: 10.48550/arxiv.1907.05600
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generative Modeling by Estimating Gradients of the Data Distribution

Abstract: We introduce a new generative model where samples are produced via Langevin dynamics using gradients of the data distribution estimated with score matching. Because gradients might be ill-defined when the data resides on low-dimensional manifolds, we perturb the data with different levels of Gaussian noise and jointly estimate the corresponding scores, i.e., the vector fields of gradients of the perturbed data distribution for all noise levels. For sampling, we propose an annealed Langevin dynamics where we us… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
85
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4

Relationship

0
9

Authors

Journals

citations
Cited by 74 publications
(97 citation statements)
references
References 34 publications
1
85
0
Order By: Relevance
“…4 were taken from. To address the problem of choosing a noise level in DSM (Saremi et al, 2018), Song & Ermon (2019) studied it with multiple noise levels by summing up the losses using a weighing scheme. See (Li et al, 2019;Chen et al, 2020;Kadkhodaie & Simoncelli, 2020;Jolicoeur-Martineau et al, 2020) in that direction.…”
Section: Resultsmentioning
confidence: 99%
“…4 were taken from. To address the problem of choosing a noise level in DSM (Saremi et al, 2018), Song & Ermon (2019) studied it with multiple noise levels by summing up the losses using a weighing scheme. See (Li et al, 2019;Chen et al, 2020;Kadkhodaie & Simoncelli, 2020;Jolicoeur-Martineau et al, 2020) in that direction.…”
Section: Resultsmentioning
confidence: 99%
“…The denoising diffusion probabilistic model (DDPM) [15] is a latent variable model where a denoising autoencoder gradually transforms Gaussian noise into real signal. Scorebased generative model [39,40] trains a neural network to predict the score function which are used to draw samples via Langevin Dynamics. Collectively, these models have demonstrated comparable or superior image quality compared to GANs while exhibiting better mode coverage and training stability.…”
Section: Related Workmentioning
confidence: 99%
“…where a parameterized model θ is estimated during the training phase. In fact, the model θ (x t , t) is just a scaled version of the score function s θ (x t , t) [38], which is the gradient of the log p θ (x t ). Once the model θ is trained, the data is sampled by the following stochastic generation step:…”
Section: Diffusion Probabilistic Modelmentioning
confidence: 99%