2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2017
DOI: 10.1109/icassp.2017.7952563
|View full text |Cite
|
Sign up to set email alerts
|

Harnessing neural networks: A random matrix approach

Abstract: This article proposes an original approach to the performance understanding of large dimensional neural networks. In this preliminary study, we study a single hidden layer feed-forward network with random input connections (also called extreme learning machine) which performs a simple regression task. By means of a new random matrix result, we prove that, as the size and cardinality of the input data and the number of neurons grow large, the network performance is asymptotically deterministic. This entails a b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2018
2018
2020
2020

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 9 publications
0
4
0
Order By: Relevance
“…Recently, the successes of deep learning along with the disqualifying complexity of studying real world problems have sparked a revived interest in the direction of random weight matrices. Recent results-without exhaustivity-were obtained on the spectrum of the Gram matrix at each layer using random matrix theory [32,33], on expressivity of deep neural networks [34], on the dynamics of propagation and learning [35][36][37][38], on the high-dimensional non-convex landscape where the learning takes place [39], or on the universal random Gaussian neural nets of [40].…”
Section: Other Related Workmentioning
confidence: 99%
“…Recently, the successes of deep learning along with the disqualifying complexity of studying real world problems have sparked a revived interest in the direction of random weight matrices. Recent results-without exhaustivity-were obtained on the spectrum of the Gram matrix at each layer using random matrix theory [32,33], on expressivity of deep neural networks [34], on the dynamics of propagation and learning [35][36][37][38], on the high-dimensional non-convex landscape where the learning takes place [39], or on the universal random Gaussian neural nets of [40].…”
Section: Other Related Workmentioning
confidence: 99%
“…In this article, following our seminal works [5,6], we propose a different angle of approach to neural network analysis. Rather than modelling a complete deep neural net, we focus here primarily on simple network structures, so far not considering backpropagation learning but accounting for nonlinearities induced when traversing a hidden layer.…”
Section: Introductionmentioning
confidence: 99%
“…In [5], we merely exploited Feature (ii) as a technical means to study the asymptotic (as n, p → ∞) performance of extreme learning machines (ELM) [7] (i.e., single hiddenlayer regression networks with no backpropagation learning), assuming a model encompassing a random connectivity matrix (which induces the concentration of the output vectors) but deterministic data. Under this model, however, while the asymptotic network training performance was readily accessible, the asymptotic generalization performance remained out of technical grasp and only a conjecture under "reasonable" yet unclear assumptions on the deterministic dataset could be proposed.…”
Section: Introductionmentioning
confidence: 99%
“…Based on similar insights that curse of dimensionality may harm RF or ELM's capabilities, in [195] they propose deep semi-random features as an alternative, which is shown to have better expressive power than RF, and better ganeralization error bound than common deep neural networks. In a different sense, in [196] the authors study the characteristics of ELM via Random Matrix Theory.…”
Section: Extreme Learning Machinementioning
confidence: 99%