1997
DOI: 10.1016/s0925-2312(97)90018-7
|View full text |Cite
|
Sign up to set email alerts
|

Advances in neural information processing systems 7

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 17 publications
(17 citation statements)
references
References 0 publications
0
17
0
Order By: Relevance
“…For instance, in many architectures the N → ∞ limit is also one in which the network is drawn from a Gaussian process (GP), where, e.g. N is the width a of a fully-connected network [39][40][41][42] or the number of channels in a CNN [43,44]. The existence of such NNGP limits is quite general [45][46][47], and allows for training with Bayesian inference [39,41].…”
Section: ∞-Nnqsmentioning
confidence: 99%
See 1 more Smart Citation
“…For instance, in many architectures the N → ∞ limit is also one in which the network is drawn from a Gaussian process (GP), where, e.g. N is the width a of a fully-connected network [39][40][41][42] or the number of channels in a CNN [43,44]. The existence of such NNGP limits is quite general [45][46][47], and allows for training with Bayesian inference [39,41].…”
Section: ∞-Nnqsmentioning
confidence: 99%
“…This ODE becomes linear and analytically solvable for a mean-squared-error loss (see the supplementary material for a review of the NTK). Similarly, in the infinite-N limit, networks are often drawn from Gaussian processes [39][40][41][42], in which case they may be trained with Bayesian inference via another deterministic constant kernel, the neural network Gaussian process (NNGP) kernel [39].…”
Section: Introductionmentioning
confidence: 99%
“…The same argument is generally valid for ensemble learning: This approach yields possible gains only if the ensemble set's predictors are different. Cios and Shields (1997) stated that the generalization error of a weighted combination of predictors in an ensemble is equal to the average error of the individual predictors minus the "disagreement" among them, which we henceforth refer to as "diversity. "…”
Section: Ensemble For Time Series Forecastingmentioning
confidence: 99%
“…Gallant described one of the first perceptron-based connectionist models where each cell i computes a single activation u i , which may be input to other cells or be an output of the network (145). Ricks et al gave one of the earliest quantum neural network models based on implementation of quantum circuitry with gates whose weights are evolved through learning using quantum search as well as piecewise weight allocation (146). In each step, we have a density operator for the qubits representing the hidden states, which can can be extracted with any output ancilliaries using a partial trace operator and fed forward to the next layer of the neural network where the unitary transformations that encapsulate the action of perceptrons can be applied.…”
Section: Quantum Reinforcement Learning and Deep Learningmentioning
confidence: 99%