1998
DOI: 10.1007/978-1-4471-1520-5_3
|View full text |Cite
|
Sign up to set email alerts
|

A Unifying View of Gradient Calculations and Learning for Locally Recurrent Neural Networks

Abstract: In this paper a critical review of gradient-based training methods for recurrent neural networks is presented including Back Propagation Through Time (BPTT), Real Time Recurrent Learning (RTRL) and several specific learning algorithms for different locally recurrent architectures. From this survey it comes out the need for a unifying view of all the specific procedures proposed for networks with local feedbacks, that keeps into account the general framework of recurrent networks learning: BPTT and RTRL. Theref… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
8
0

Year Published

2002
2002
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(8 citation statements)
references
References 23 publications
(49 reference statements)
0
8
0
Order By: Relevance
“…After using these algorithms with some static problems, with good performances, we have applied them to training locally recurrent neural networks, as explained in the next section. [6,7], and which has good performances.…”
Section: R (W>= P L (23mentioning
confidence: 98%
See 1 more Smart Citation
“…After using these algorithms with some static problems, with good performances, we have applied them to training locally recurrent neural networks, as explained in the next section. [6,7], and which has good performances.…”
Section: R (W>= P L (23mentioning
confidence: 98%
“…With the same notation used in [6,7,8] As stated above, applying the tW0 previous techniques to calculate the products H(%)p , we obtain two new algorithms based on the conjugate gradient method.…”
Section: Scg-r and Scgu Algorithms For Iir-mlp Neural Networkmentioning
confidence: 99%
“…The obtained gradient information can be used both for parameter sensitivity computation of systems and for training of adaptive systems. The system can be any causal, in general non-linear and time-variant, dynamical system represented by a SFG, in particular any feedforward (static), time delay or recurrent neural networks [2,3,9]. In this work we use discrete time notation, but the same theory holds for the continuous time case.…”
Section: Introductionmentioning
confidence: 99%
“…On-line training of adaptive systems is very important in real applications such as Digital Signal Processing, system identification and control, channel equalization and predistortion, since it allows the model to follow time-varying features of the real system [2,3]. By on-line learning in contrast to batch learning this is provided with no interruption of the operation of the system (such as a communication channel).…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation