2022
DOI: 10.1007/s42519-021-00229-5
|View full text |Cite
|
Sign up to set email alerts
|

Function Approximation by Deep Neural Networks with Parameters $$\{0,\pm \frac{1}{2}, \pm 1, 2\}$$

Abstract: In this paper, it is shown that $$C_\beta $$ C β -smooth functions can be approximated by deep neural networks with ReLU activation function and with parameters $$\{0,\pm \frac{1}{2}, \pm 1, 2\}$$ { 0 , ± 1 2 , … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
1
1

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 7 publications
0
4
0
Order By: Relevance
“…Approximations of β-Hölder functions belonging to the ball [9]. Using that result and substituting the interval [−1, 1] by the finite set of weights {0, ± 1 2 , ±1, 2}, the following theorem is given in [2]:…”
Section: Relu Networkmentioning
confidence: 99%
“…Approximations of β-Hölder functions belonging to the ball [9]. Using that result and substituting the interval [−1, 1] by the finite set of weights {0, ± 1 2 , ±1, 2}, the following theorem is given in [2]:…”
Section: Relu Networkmentioning
confidence: 99%
“…(i) deriving prediction rates of the empirical risk minimizers (1) or ( 2); (ii) finding an optimization algorithm that identifies the corresponding empirical risk minimizers. Convergence rates of empirical risk minimizers (ERM) over the classes of deep ReLU networks are studied in [4], [13], [15] and [18]. In [4] it is shown that the ERM of the form (1), with W n being the set of weight vectors with coordinates {0, ±1/2, ±1, 2}, attains, up to logarithmic factors, the minimax rates of prediction of β-smooth functions.…”
Section: Introductionmentioning
confidence: 99%
“…Convergence rates of empirical risk minimizers (ERM) over the classes of deep ReLU networks are studied in [4], [13], [15] and [18]. In [4] it is shown that the ERM of the form (1), with W n being the set of weight vectors with coordinates {0, ±1/2, ±1, 2}, attains, up to logarithmic factors, the minimax rates of prediction of β-smooth functions. The finiteness of W n guarantees that in this case the ERM can be found within finitely many steps.…”
Section: Introductionmentioning
confidence: 99%
“…We will consider the family of activation functions A = {σ, ⌊•⌋}, where the role of the activation σ is to guarantee that the conditions of the Kronecker's theorem are satisfied and that it gives a small range for the integer multipliers. Having this range we then bound the entropy of approximant networks and use this bound to get for β-Hölder continuous regression functions a convergence rate of order n −2β 2β+d log 2 n. Note that our approach is in some sense opposite to the one given in [1], where approximations by deep networks with weights {0, ± 1 2 , ±1, 2} are considered: in one case we fix a finite set of weights and adjust the network architecture and in the other case we fix the network architecture and adjust the integer weights to attain a certain approximation rate.…”
mentioning
confidence: 99%