2020
DOI: 10.1088/1742-5468/abcaee
|View full text |Cite
|
Sign up to set email alerts
|

Conformal symplectic and relativistic optimization

Abstract: Arguably, the two most popular accelerated or momentum-based optimization methods in machine learning are Nesterov’s accelerated gradient and Polyaks’s heavy ball, both corresponding to different discretizations of a particular second order differential equation with friction. Such connections with continuous-time dynamical systems have been instrumental in demystifying acceleration phenomena in optimization. Here we study structure-preserving discretizations for a certain class of dissipative (conformal) Hami… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
53
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 25 publications
(53 citation statements)
references
References 27 publications
0
53
0
Order By: Relevance
“…Then, using a numerical integration method that preserves the property of Hamiltonian energy dissipation (such a numerical method is called conformal symplectic scheme [13]), we get an approximation of (θ * , 0), where (θ * , 0) is an equilibrium of (6.4) and θ * a stationary point of E [99]. In recent work, Frana et al [53] take advantage of the connection between optimisation schemes for E and conformal symplectic Hamiltonian schemes for H = T + E for the design of new optimisation approaches for E -similar to the connections we have seen before, e.g. between Adam and conformal Hamiltonian descent.…”
Section: Conformal Hamiltonian Systemsmentioning
confidence: 99%
“…Then, using a numerical integration method that preserves the property of Hamiltonian energy dissipation (such a numerical method is called conformal symplectic scheme [13]), we get an approximation of (θ * , 0), where (θ * , 0) is an equilibrium of (6.4) and θ * a stationary point of E [99]. In recent work, Frana et al [53] take advantage of the connection between optimisation schemes for E and conformal symplectic Hamiltonian schemes for H = T + E for the design of new optimisation approaches for E -similar to the connections we have seen before, e.g. between Adam and conformal Hamiltonian descent.…”
Section: Conformal Hamiltonian Systemsmentioning
confidence: 99%
“…We emphasize here that since absolute stable theory is corresponding to accumulated error, it can be widely used in iterative algorithms, although this concept was orginally proposed for numerical integrators. Similar analyses are applied to different optimization methods in Su et al [2016] and França et al [2020b].…”
Section: Better Stability Of Sagmentioning
confidence: 99%
“…[113] applied Runge-Kutta integration to an inertial gradient system without Hessian-driven damping [110] and showed that the resulting algorithm is faster than NAG when the objective function is sufficiently smooth and when the order of the integrator is sufficiently large. [78] and [60] both considered conformal Hamiltonian systems and showed that the resulting discretetime algorithm achieves fast convergence under certain smoothness conditions. Very recently, [103] have rigorously justified the use of symplectic Euler integrators compared to explicit and implicit Euler integration, which was further studied by [59,87].…”
Section: Introductionmentioning
confidence: 99%