2020
DOI: 10.1016/j.neunet.2019.12.014
|View full text |Cite
|
Sign up to set email alerts
|

On the minimax optimality and superiority of deep neural network learning over sparse parameter spaces

Abstract: Deep learning has been applied to various tasks in the field of machine learning and has shown superiority to other common procedures such as kernel methods. To provide a better theoretical understanding of the reasons for its success, we discuss the performance of deep learning and other methods on a nonparametric regression problem with a Gaussian noise. Whereas existing theoretical studies of deep learning have been based mainly on mathematical theories of well-known function classes such as Hölder and Beso… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
31
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 29 publications
(31 citation statements)
references
References 27 publications
0
31
0
Order By: Relevance
“…In contrast, ML, deep learning (DL), and AI excel at automatic pattern recognition from large amounts of biomedical image data. In particular, machine learning and deep learning algorithms (e.g., support vector machine, neural network, and convolutional neural network) have achieved impressive results in biomedical image classification [14][15][16][17][18][19][20][21][22][23]. Classification helps to organize biomedical image databases into image categories before diagnostics [24][25][26][27][28][29][30].…”
Section: A Survey Of Biomedical Imagementioning
confidence: 99%
See 1 more Smart Citation
“…In contrast, ML, deep learning (DL), and AI excel at automatic pattern recognition from large amounts of biomedical image data. In particular, machine learning and deep learning algorithms (e.g., support vector machine, neural network, and convolutional neural network) have achieved impressive results in biomedical image classification [14][15][16][17][18][19][20][21][22][23]. Classification helps to organize biomedical image databases into image categories before diagnostics [24][25][26][27][28][29][30].…”
Section: A Survey Of Biomedical Imagementioning
confidence: 99%
“…SVM is a margin-based classifier that achieves superior classification performance compared to other algorithms when the amount of dataset training is medium [34,51,60]. DL techniques are conquering the prevailing traditional approaches of the neural network; when it comes to the huge amount of dataset, applications requiring complex functions demanding increase accuracy with lower time complexities [22,66,67]. DL particularly CNN has shown an intrinsic ability to automatically extract the high-level representations from big data [36].…”
Section: Machine Learning Algorithm (Svm or Cnn)mentioning
confidence: 99%
“…The idea that machine learning methods should be able learn both smooth as well as non-smooth functions has recently received attention also in other areas of machine learning. For instance, Imaizumi and Fukumizu [24] and Hayakawa and Suzuki [20] argue that one of the reasons for the superior predictive accuracy of deep neural networks, over e.g. kernel methods, is their ability to also learn non-smooth functions.…”
Section: Related Workmentioning
confidence: 99%
“…In Section 6, KBW divide statistical procedures into structure-driven and methods-driven but also acknowledge that the boundary between these two categories is blurry. For example, even for the poster child of the methods-driven tools -deep neural networks -one common research direction is to prove some form of optimality or robustness under some assumptions, often quantified by smoothness, sparsity or other related complexity measures such as metric entropy (Schmidt-Hieber, 2020, Hayakawa and Suzuki, 2020, Barron and Klusowski, 2018.…”
Section: A Brief Introduction To Higher Order Influence Functionsmentioning
confidence: 99%
“…In Section 6, KBW divide statistical procedures into structure-driven and methods-driven but also acknowledge that the boundary between these two categories is blurry. For example, even for the poster child of the methods-driven tools -deep neural networks -one common research direction is to prove some form of optimality or robustness under some assumptions, often quantified by smoothness, sparsity or other related complexity measures such as metric entropy (Schmidt-Hieber, 2020, Hayakawa and Suzuki, 2020, Barron and Klusowski, 2018.The discussants then state that higher order influence function (HOIF) based methods are 'structure-driven' because 'they typically rely on carefully constructed series estimates' and achieve 'better performance over appropriate Hölder spaces potentially at the expense of being more structure driven.' This statement misunderstands the motivation and goals of HOIF estimation.…”
mentioning
confidence: 99%