2020
DOI: 10.1038/s41598-020-76764-1
|View full text |Cite
|
Sign up to set email alerts
|

Power-law scaling to assist with key challenges in artificial intelligence

Abstract: Power-law scaling, a central concept in critical phenomena, is found to be useful in deep learning, where optimized test errors on handwritten digit examples converge as a power-law to zero with database size. For rapid decision making with one training epoch, each example is presented only once to the trained network, the power-law exponent increased with the number of hidden layers. For the largest dataset, the obtained test error was estimated to be in the proximity of state-of-the-art algorithms for large … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 14 publications
(7 citation statements)
references
References 30 publications
0
7
0
Order By: Relevance
“…Models that recognize objects from images require about a 500-fold increase in resources to only double their intelligence [4]. Other studies reached similar conclusions [5,6] (see [7] for a review of how machine learning engineers and data scientists cope in practice with these limitations of deep learning models).…”
Section: Introductionmentioning
confidence: 70%
“…Models that recognize objects from images require about a 500-fold increase in resources to only double their intelligence [4]. Other studies reached similar conclusions [5,6] (see [7] for a review of how machine learning engineers and data scientists cope in practice with these limitations of deep learning models).…”
Section: Introductionmentioning
confidence: 70%
“…3). We note that the presented power law as a function of the depth of the architecture differs from the power law behavior for SRs as a function of the dataset size [29][30][31][32][33] .…”
Section: Discussionmentioning
confidence: 99%
“…Another possible mechanism is the addition of a super-linear number of cross-weights to the filters. This represents a biological realization because cross-weights result as a byproduct of dendritic nonlinear amplification 17,29,34,35 . Nevertheless, these possible enhanced ρ mechanisms significantly increase computational complexity and are mentioned for their potential biological relevance, limited number of layers, and the natural emergence of many cross-weights.…”
Section: Discussionmentioning
confidence: 99%
“…1. This relation advises on the suitable rescaling of the dataset size (M r −2 ), as the dataset quality is impaired (r → 0), in order to preserve network's abilities; note that power-law scalings were already evidenced in the machine-learning context, see, e.g., [18]. To achieve a quantitative picture and control of the network behavior, we work out a statistical-mechanics investigation and we start by introducing the Boltzmann-Gibbs measure for the system,…”
mentioning
confidence: 99%