2019
DOI: 10.48550/arxiv.1901.03611
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Benefits of Over-parameterization at Initialization in Deep ReLU Networks

Abstract: It has been noted in existing literature that over-parameterization in ReLU networks generally improves performance. While there could be several factors involved behind this, we prove some desirable theoretical properties at initialization which may be enjoyed by ReLU networks. Specifically, it is known that He initialization in deep ReLU networks asymptotically preserves variance of activations in the forward pass and variance of gradients in the backward pass for infinitely wide networks, thus preserving th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
21
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
4

Relationship

0
8

Authors

Journals

citations
Cited by 14 publications
(22 citation statements)
references
References 9 publications
1
21
0
Order By: Relevance
“…Theorem 4.1 is new in its own right and improves on a previous result by Giryes et al [GSB16], see also [GSB20]. It is also closely related to work of Arpit and Bengio [AB19], who investigated the capability of a random ReLU layer as in Definition 1.2, but with bias b = 0, to preserve Euclidean norms.…”
Section: Random Embeddingssupporting
confidence: 66%
“…Theorem 4.1 is new in its own right and improves on a previous result by Giryes et al [GSB16], see also [GSB20]. It is also closely related to work of Arpit and Bengio [AB19], who investigated the capability of a random ReLU layer as in Definition 1.2, but with bias b = 0, to preserve Euclidean norms.…”
Section: Random Embeddingssupporting
confidence: 66%
“…Altering the derivation including rectified linear units (ReLU) led to the "He" initialization (He et al, 2015). "He" initialization seems also to be favorably related to over-parameterization of networks (Arpit & Bengio, 2019). Similar to these early works, more recent work has also considered activation scale (Hanin & Rolnick, 2018) and gradients (Balduzzi et al, 2017) as well as dynamical isometry properties (Saxe et al, 2013;Yang & Schoenholz, 2017).…”
Section: Amount Of Datamentioning
confidence: 94%
“…1 (a), it is clear that overparametrization increases the number of subdivision lines and the chance to well position some of them. In addition, overparametrization has also been used favorably as a way to help gradient descent based techniques facilitating optimization (Arpit & Bengio, 2019); lastly, overparametrized also positions the initial parameters close to good local minima (Allen-Zhu et al, 2019;Zou & Gu, 2019;Kawaguchi et al, 2019) reducing the amount of updated needed during training. We formalize those points in the remark below.…”
Section: The Initialization Dilemma and The Importance Of Overparamet...mentioning
confidence: 99%