2019
DOI: 10.48550/arxiv.1910.02008
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Nonasymptotic estimates for Stochastic Gradient Langevin Dynamics under local conditions in nonconvex optimization

Abstract: Within the context of empirical risk minimization, see Raginsky, Rakhlin, and Telgarsky (2017), we are concerned with a non-asymptotic analysis of sampling algorithms used in optimization. In particular, we obtain non-asymptotic error bounds for a popular class of algorithms called Stochastic Gradient Langevin Dynamics (SGLD). These results are derived in appropriate Wasserstein distances in the absence of log-concavity of the target distribution. More precisely, the stochastic gradient H(θ, x) is assumed to b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
23
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 14 publications
(23 citation statements)
references
References 20 publications
0
23
0
Order By: Relevance
“…In this section, we start with placing assumptions on stochastic gradients H(θ, ε) as defined in (8). We note that these assumptions are the most relaxed conditions to prove the convergence of Langevin dynamics to this date, see, e.g., Zhang et al (2019), Chau et al (2021). We first need to assume that sufficient moments of the distribution r ε exists.…”
Section: Convergence Rates Of Solaismentioning
confidence: 98%
See 3 more Smart Citations
“…In this section, we start with placing assumptions on stochastic gradients H(θ, ε) as defined in (8). We note that these assumptions are the most relaxed conditions to prove the convergence of Langevin dynamics to this date, see, e.g., Zhang et al (2019), Chau et al (2021). We first need to assume that sufficient moments of the distribution r ε exists.…”
Section: Convergence Rates Of Solaismentioning
confidence: 98%
“…This will lead to a global optimizer θ ⋆ which will give the best possible proposal in terms of minimizing the MSE of the importance sampler. We use stochastic gradient Langevin dynamics (SGLD) (Zhang et al, 2019) and its underdamped counterpart stochastic gradient Hamiltonian Monte Carlo (SGHMC) (Akyildiz and Sabanis, 2020) for global optimization. We summarize the algorithms in the next section.…”
Section: Adaptation As Global Nonconvex Optimizationmentioning
confidence: 99%
See 2 more Smart Citations
“…Remark 3.4. Langevin sampling method has been studied under the (strongly) convex potential assumption [15,16,17,11,12,8,13]; the dissipativity condition for the drift term [34,33,40]; the local convexity condition for the potential function outside a ball [17,9,30,3]. However, these conditions may not hold for models with multiple modes, for example, Gaussian mixtures, where their potentials are not convex and the log Sobolev inequality may not be satisfied.…”
mentioning
confidence: 99%