2021
DOI: 10.48550/arxiv.2107.12530
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Convergence of Deep ReLU Networks

Abstract: We explore convergence of deep neural networks with the popular ReLU activation function, as the depth of the networks tends to infinity. To this end, we introduce the notion of activation domains and activation matrices of a ReLU network. By replacing applications of the ReLU activation function by multiplications with activation matrices on activation domains, we obtain an explicit expression of the ReLU network. We then identify the convergence of the ReLU networks as convergence of a class of infinite prod… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

3
44
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

2
3

Authors

Journals

citations
Cited by 6 publications
(48 citation statements)
references
References 18 publications
3
44
1
Order By: Relevance
“…Having the above matrix form, conditions on the masks and bias vectors that ensure convergence of the convolutional neural network (2.1) as the depth n tends to infinity are now reformulated as conditions on the weight matrices and the bias vectors. Convergence of deep ReLU neural networks with a fixed width was recently studied in [21]. The matrix form for CNNs differs from that for DNNs considered in [21] in that the widths in CNNs are increasing while those in [21] are fixed.…”
Section: Convolutional Neural Networkmentioning
confidence: 99%
See 4 more Smart Citations
“…Having the above matrix form, conditions on the masks and bias vectors that ensure convergence of the convolutional neural network (2.1) as the depth n tends to infinity are now reformulated as conditions on the weight matrices and the bias vectors. Convergence of deep ReLU neural networks with a fixed width was recently studied in [21]. The matrix form for CNNs differs from that for DNNs considered in [21] in that the widths in CNNs are increasing while those in [21] are fixed.…”
Section: Convolutional Neural Networkmentioning
confidence: 99%
“…In particular, the approximation and expressive powers of deep neural networks have been a focus. A brief review of the literature for this aspect was contained in the introduction of [21]. We also refer readers to the two recent surveys [4,6] for further details.…”
Section: Introductionmentioning
confidence: 99%
See 3 more Smart Citations