2019
DOI: 10.48550/arxiv.1910.12947
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Generalization Bounds of a Family of Recurrent Neural Networks

Abstract: Recurrent Neural Networks (RNNs) have been widely applied to sequential data analysis. Due to their complicated modeling structures, however, the theory behind is still largely missing. To connect theory and practice, we study the generalization properties of vanilla RNNs as well as their variants, including Minimal Gated Unit (MGU), Long Short Term Memory (LSTM), and Convolutional (Conv) RNNs. Specifically, our theory is established under the PAC-Learning framework. The generalization bound is presented in te… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
11
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 8 publications
(12 citation statements)
references
References 29 publications
1
11
0
Order By: Relevance
“…Next, we study the generalization ability of GNNs via Rademacher bounds, focusing on binary classification. We generalize the previous results on the complexity of feedforward networks (Bartlett et al, 2017;Neyshabur et al, 2018) and RNNs (Chen et al, 2019a) in mainly three ways. First, we process graphs unlike sequences in RNNs, or instances restricted to the input layer in feedforward networks.…”
Section: Generalization Bounds For Gnnssupporting
confidence: 79%
See 3 more Smart Citations
“…Next, we study the generalization ability of GNNs via Rademacher bounds, focusing on binary classification. We generalize the previous results on the complexity of feedforward networks (Bartlett et al, 2017;Neyshabur et al, 2018) and RNNs (Chen et al, 2019a) in mainly three ways. First, we process graphs unlike sequences in RNNs, or instances restricted to the input layer in feedforward networks.…”
Section: Generalization Bounds For Gnnssupporting
confidence: 79%
“…We also mention the corresponding bounds for RNN on a sequence of length L when the spectral norm of recurrent weights in RNN is respectively less than, equal to, or greater than 1 (note that we renamed some parameters from Chen et al (2019a) for notational consistency). Our analysis implies GNNs have essentially the same dependence on dimension r, C GNN (ours) RNN (Chen et al (2019a))…”
Section: Generalization Bound For Gnnsmentioning
confidence: 69%
See 2 more Smart Citations
“…Most recently, Sato et al (2020) provides PAC learning-style bounds on the node embedding and gradient estimation for SGCNs training. Another direction of theoretical research focuses on analyzing the expressive power of GCN (Garg et al, 2020;Chen et al, 2019;, which is not the focus of this paper and omitted for brevity.…”
mentioning
confidence: 99%