2022
DOI: 10.1609/aaai.v36i6.20582
|View full text |Cite
|
Sign up to set email alerts
|

KAM Theory Meets Statistical Learning Theory: Hamiltonian Neural Networks with Non-zero Training Loss

Abstract: Many physical phenomena are described by Hamiltonian mechanics using an energy function (Hamiltonian). Recently, the Hamiltonian neural network, which approximates the Hamiltonian by a neural network, and its extensions have attracted much attention. This is a very powerful method, but theoretical studies are limited. In this study, by combining the statistical learning theory and KAM theory, we provide a theoretical analysis of the behavior of Hamiltonian neural networks when the learning error is not complet… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
3
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 15 publications
0
3
0
Order By: Relevance
“…[16] further improves the model by introducing a learnable 1 × 1 invertible convolution to replace the fixed permutation of channel dimension used in [15]. In connection with NODE, [17] leverages INNs to map the underlying vector field of an ODE system to a base vector field, and others [18,19] have attempted to include INNs in Hamiltonian systems.…”
Section: Related Workmentioning
confidence: 99%
“…[16] further improves the model by introducing a learnable 1 × 1 invertible convolution to replace the fixed permutation of channel dimension used in [15]. In connection with NODE, [17] leverages INNs to map the underlying vector field of an ODE system to a base vector field, and others [18,19] have attempted to include INNs in Hamiltonian systems.…”
Section: Related Workmentioning
confidence: 99%
“…Recent developments in automatic differentiation have made it possible to easily train neural networks that have special architectures [29], [30]. Following that, numerous studies have investigated new architectures for specific subclasses of continuous-time dynamical systems, including Hamiltonian systems on canonical coordinates [31], on arbitrary coordinates [32], and with constraints [33], [34], [35], Euler-Lagrange systems [36], and port-Hamiltonian systems [37]. These systems are associated with special geometric structures (e.g., symplectic structure), and the neural networks will be able to learn these systems with high accuracy if their architectures represent such structures well.…”
mentioning
confidence: 99%
“…The proposed method obtains the gradient from each step as a numerical integration and is, thus, more robust to rounding errors. 4) Compatible with any ODE systems-while this advantage applies to comparison methods, we emphasize that the proposed method is compatible with any systems described by ODEs [31], [32], [33], [34], [35], [36], [37], [38], [39], [40], [41], [43].…”
mentioning
confidence: 99%