“…Specifically, we use stochastic gradient Langevin dynamics (SGLD) (Welling and Teh, 2011) and its underdamped counterpart, stochastic gradient Hamiltonian Monte Carlo (SGHMC) (Chen et al, 2014), for non-convex optimization. Recently, global convergence of these algorithms for nonconvex optimization were shown in several works, see, e.g., Raginsky et al (2017), Xu et al (2018), Erdogdu et al (2018, Zhang et al (2019), Akyildiz and Sabanis (2020), Gao et al (2021), , . We leverage these results for proving that optimizing a general non-convex χ 2 -divergence leads to a global convergence result for the resulting AIS schemes.…”