2020
DOI: 10.1088/2632-2153/aba8e7
|View full text |Cite
|
Sign up to set email alerts
|

Fractional deep neural network via constrained optimization

Abstract: This paper introduces a novel algorithmic framework for a deep neural network (DNN), which in a mathematically rigorous manner, allows us to incorporate history (or memory) into the network—it ensures all layers are connected to one another. This DNN, called Fractional-DNN, can be viewed as a time-discretization of a fractional in time non-linear ordinary differential equation (ODE). The learning problem then is a minimization problem subject to that fractional ODE as constraints. We emphasize that an analogy … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
59
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 20 publications
(59 citation statements)
references
References 43 publications
0
59
0
Order By: Relevance
“…Besides, Antil in Ref. [50] proposed a deep FANN for the time-discretization of a fractional in time nonlinear ordinary DE, employing Caputo derivative. The fractional ordinary DE was minimized by the learning algorithm, solving several issues, such as network instability, vanishing, exploding gradients, long training times, and inability to approximate non-smooth functions.…”
Section: Gradient Descent Algorithm (Gd)mentioning
confidence: 99%
“…Besides, Antil in Ref. [50] proposed a deep FANN for the time-discretization of a fractional in time nonlinear ordinary DE, employing Caputo derivative. The fractional ordinary DE was minimized by the learning algorithm, solving several issues, such as network instability, vanishing, exploding gradients, long training times, and inability to approximate non-smooth functions.…”
Section: Gradient Descent Algorithm (Gd)mentioning
confidence: 99%
“…Deep learning (DL), a new Artificial Intelligence (AI) trend that uses multi-layer perception network [8], has received increasing attention from researchers and has been widely applied to numerous real-world applications and across many fields [2,9,10,11,12,13,14,15]. Deep learning is able to effectively capture the non-linear and non-trivial user-item relationships, and it enables the codification of more complex abstractions as data representations in the higher layers.…”
Section: Deep Learningmentioning
confidence: 99%
“…The DNNs used in this paper have been motivated by the Residual Neural Network (ResNet) architecture. ResNets have been introduced in [9, 24,2] in the context of data/image classification, see also [1] for parameterized PDEs and [7] where the (related) so-called Neural ODE Nets [4] have been used to solve stiff ODEs. The ResNet architecture is known to overcome the vanishing gradient problem, which has been further analyzed using fractional order derivatives in [2].…”
Section: Introductionmentioning
confidence: 99%
“…ResNets have been introduced in [9, 24,2] in the context of data/image classification, see also [1] for parameterized PDEs and [7] where the (related) so-called Neural ODE Nets [4] have been used to solve stiff ODEs. The ResNet architecture is known to overcome the vanishing gradient problem, which has been further analyzed using fractional order derivatives in [2]. The key feature of a ResNet is that in the continuum limit, it becomes an optimization problem constrained by an ODE.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation