2019
DOI: 10.1038/s41598-019-50158-4
|View full text |Cite
|
Sign up to set email alerts
|

Echo State Networks with Self-Normalizing Activations on the Hyper-Sphere

Abstract: Among the various architectures of Recurrent Neural Networks, Echo State Networks (ESNs) emerged due to their simplified and inexpensive training procedure. These networks are known to be sensitive to the setting of hyper-parameters, which critically affect their behavior. Results show that their performance is usually maximized in a narrow region of hyper-parameter space called edge of criticality. Finding such a region requires searching in hyper-parameter space in a sensible way: hyper-parameter configurati… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
15
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 23 publications
(16 citation statements)
references
References 41 publications
1
15
0
Order By: Relevance
“…The commonlyused tanh activation offers good accuracy [13,18] for the systems studied here, as discussed in the results sections 3.1 and 3.2. While different activation functions have been proposed [19], it is beyond the scope of the present work to study the effect of activation functions on the echo state network accuracy. In the conventional ESN approach (Fig.…”
Section: Methodsmentioning
confidence: 99%
“…The commonlyused tanh activation offers good accuracy [13,18] for the systems studied here, as discussed in the results sections 3.1 and 3.2. While different activation functions have been proposed [19], it is beyond the scope of the present work to study the effect of activation functions on the echo state network accuracy. In the conventional ESN approach (Fig.…”
Section: Methodsmentioning
confidence: 99%
“…The ESN is a promising type of RNNs due to its relaxed training complexity and its ability to preserve the temporal features from different signals over time [21], [37]- [40]. The ESNs are in the reservoir computing (RC) category because in the ESNs only the output weights are trainable.…”
Section: Echo State Networkmentioning
confidence: 99%
“…Finally, W in ∈ R Nr×ni is the input weight matrix, µ is the leaking rate parameter, and W o ∈ R no×Nr is the output weight matrix which is the only one that is trainable using a regression technique. This training phase in ESN does not affect the dynamics of the system, which makes it possible to operate with the same reservoir for different tasks [37]. A schematic representation of a leaky-ESN, including the sequential input, dynamic, static, and output layers, as depicted in Fig.…”
Section: Echo State Networkmentioning
confidence: 99%
“…Other hyper-parameters refer to the input and output scaling factors, the sparsity degree of W and the input bias. Moreover, the update equation ( 1) is usually chosen to be non-linear, i.e., x k = φ(W x k−1 + wu k ) and different choices of the nonlinear transfer function φ can be explored [42].…”
Section: Reservoir Computingmentioning
confidence: 99%