2019
DOI: 10.1142/s0219530519400074
|View full text |Cite
|
Sign up to set email alerts
|

Deep neural networks for rotation-invariance approximation and learning

Abstract: Based on the tree architecture, the objective of this paper is to design deep neural networks with two or more hidden layers (called deep nets) for realization of radial functions so as to enable rotational invariance for near-optimal function approximation in an arbitrarily high dimensional Euclidian space. It is shown that deep nets have much better performance than shallow nets (with only one hidden layer) in terms of approximation accuracy and learning capabilities. In particular, for learning radial funct… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
59
0
1

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 37 publications
(62 citation statements)
references
References 38 publications
2
59
0
1
Order By: Relevance
“…The reason is that we focus on all activation functions satisfying (1) rather than piecewise activation functions. It should be also mentioned that similar covering number estimates for deep nets with tree structures has been studied in [6], [14], [25]. We highlight that different structures yield essentially non-trivial approaches.…”
Section: Covering Number Estimatessupporting
confidence: 53%
See 3 more Smart Citations
“…The reason is that we focus on all activation functions satisfying (1) rather than piecewise activation functions. It should be also mentioned that similar covering number estimates for deep nets with tree structures has been studied in [6], [14], [25]. We highlight that different structures yield essentially non-trivial approaches.…”
Section: Covering Number Estimatessupporting
confidence: 53%
“…We also exhibit that for some simple data features like the smoothness, deep nets performs essentially similar as shallow nets. 6 APPENDIX A: PROOF OF THEOREM 1 For 1 ≤ ℓ ≤ L, let W * F ℓ,w be the set of d ℓ × d ℓ−1 matrices with fixed structures and total F ℓ,w free parameters and B * F ℓ,b be the set of F ℓ,b -dimensional vectors with fixed structures and total F ℓ,b free parameters. Denote…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Kernel methods have been developed as powerful tools for machine learning, statistical analysis and probability numeral calculations [16], [8]. In the recent years, scientists have paid large attention to kernel methods to study various learning frameworks, such as extreme learning [19], deep learning [2], Bayesian learning [10] and others [14], [4], [15]. We attempt to extend the kernel method to find a suitable loss function for CNN problem.…”
mentioning
confidence: 99%