2020
DOI: 10.48550/arxiv.2012.04115
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generalization bounds for deep learning

Guillermo Valle-Pérez,
Ard A. Louis

Abstract: Generalization in deep learning has been the topic of much recent theoretical and empirical research. Here we introduce desiderata for techniques that predict generalization errors for deep learning models in supervised learning. Such predictions should 1) scale correctly with data complexity; 2) scale correctly with training set size; 3) capture differences between architectures; 4) capture differences between optimization algorithms; 5) be quantitatively not too far from the true error (in particular, be non… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 52 publications
0
4
0
Order By: Relevance
“…For example, the complexity of the hypothesis space bound increases as the number of parameters becomes larger and the stability-based bound grows with respect to the iterations of optimization. To handle the problems and further improve the generalization bound of GNNs, researchers can leverage recent advances in deep learning theory such as local Rademacher complexity [178], marginal-likelihood PAC-Bayes [179] and H-consistency [180]. Besides, it is observed that existing generalization bounds often heavily rely on the number of nodes and maximum node degree as the graph-related term in their final expressions, which is too coarse-grained to capture the complex graph structure information.…”
Section: Discussionmentioning
confidence: 99%
“…For example, the complexity of the hypothesis space bound increases as the number of parameters becomes larger and the stability-based bound grows with respect to the iterations of optimization. To handle the problems and further improve the generalization bound of GNNs, researchers can leverage recent advances in deep learning theory such as local Rademacher complexity [178], marginal-likelihood PAC-Bayes [179] and H-consistency [180]. Besides, it is observed that existing generalization bounds often heavily rely on the number of nodes and maximum node degree as the graph-related term in their final expressions, which is too coarse-grained to capture the complex graph structure information.…”
Section: Discussionmentioning
confidence: 99%
“…A comprehensive review by Valle-Pérez and Louis (2020) shows that, most types of generalization bounds are characterized by the norm of network weights or other related quantities such as the Lipschitz constant of the network; some examples include the margin-based bounds (Bartlett et al, 2017), the sensitivity-based bounds (Neyshabur et al, 2017), the NTK-based bounds (Cao and Gu, 2019) or the compression-based bounds (Li et al, 2020). They are model-agnostic in the sense that, only a generic functional form of the network is required and the weights are then implicitly regularized by some learning algorithms.…”
Section: Comparison With Other Existing Workmentioning
confidence: 99%
“…Generalization in reinforcement learning Generalizing a model's predictions across a variety of unseen, high-dimensional inputs has been extensively studied in the static supervised learning setting [Bartlett, 1998, Triantafillou et al, 2019, Valle-Pérez and Louis, 2020, Liu et al, 2020b. Gener-…”
Section: Related Workmentioning
confidence: 99%