2019
DOI: 10.48550/arxiv.1908.10920
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Deep Learning Theory Review: An Optimal Control and Dynamical Systems Perspective

Guan-Horng Liu,
Evangelos A. Theodorou

Abstract: Attempts from different disciplines to provide a fundamental understanding of deep learning have advanced rapidly in recent years, yet a unified framework remains relatively limited. In this article, we provide one possible way to align existing branches of deep learning theory through the lens of dynamical system and optimal control. By viewing deep neural networks as discrete-time nonlinear dynamical systems, we can analyze how information propagates through layers using mean field theory. When optimization … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
22
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
6
4

Relationship

1
9

Authors

Journals

citations
Cited by 18 publications
(22 citation statements)
references
References 68 publications
(161 reference statements)
0
22
0
Order By: Relevance
“…In this context, the training of the network can be interpreted as a large optimal control problem, an insight that was proposed independently by Weinan E [31] and Haber-Ruthotto [42]. Later on, this dynamical approach has been greatly popularized in the machine learning community under the name of NeurODE by Chen et al [27], see also [52]. The formulation starts by re-interpreting the iteration (1.2) as a discrete-time Euler approximation [9] of the following dynamical system Ẋt = F(t, X t , θ t ) ,…”
Section: Neurodes and Stochastic Optimal Controlmentioning
confidence: 99%
“…In this context, the training of the network can be interpreted as a large optimal control problem, an insight that was proposed independently by Weinan E [31] and Haber-Ruthotto [42]. Later on, this dynamical approach has been greatly popularized in the machine learning community under the name of NeurODE by Chen et al [27], see also [52]. The formulation starts by re-interpreting the iteration (1.2) as a discrete-time Euler approximation [9] of the following dynamical system Ẋt = F(t, X t , θ t ) ,…”
Section: Neurodes and Stochastic Optimal Controlmentioning
confidence: 99%
“…Note that this FDR relation for GANs training is analogous to that for stochastic gradient descent algorithm on a pure minimization problem in [Yaida, 2019] and [Liu and Theodorou, 2019]. This FDR relation in GANs reveals the crucial difference between GANs training of discriminator and generator networks versus training of two independent neural networks.…”
Section: Dynamics Of Training Loss and Fdrmentioning
confidence: 75%
“…A practical training algorithm for deep learning can then be obtained by approximating the optimal solution based on the PMP via various numerical approximation (Chernousko and Lyubushin, 1982;Krylov and Chernous ḱo, 1972;LeCun, 1988). See, e.g., Benning et al (2019);E (2017);E et al (2019); Han and E (2016); Li and Hao (2018); Li et al (2017); Liu and Markowich (2019); Zhang et al (2019) for more details of these works, and see Liu and Theodorou (2019) for a recent survey on the connection between deep learning and optimal control theory.…”
Section: Background and Related Workmentioning
confidence: 99%