1992
DOI: 10.1109/72.105429
|View full text |Cite
|
Sign up to set email alerts
|

Weight perturbation: an optimal architecture and learning technique for analog VLSI feedforward and recurrent multilayer networks

Abstract: Previous work on analog VLSI implementation of multilayer perceptrons with on-chip learning has mainly targeted the implementation of algorithms such as back-propagation. Although back-propagation is efficient, its implementation in analog VLSI requires excessive computational hardware. It is shown that using gradient descent with direct approximation of the gradient instead of back-propagation is more economical for parallel analog implementations. It is shown that this technique (which is called ;weight pert… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
36
0

Year Published

1997
1997
2014
2014

Publication Types

Select...
5
5

Relationship

0
10

Authors

Journals

citations
Cited by 193 publications
(36 citation statements)
references
References 11 publications
0
36
0
Order By: Relevance
“…Several studies have been performed with optimization routines guiding systematic experiments or computational fluid dynamics (CFD) trials (Kern & Koumoutsakos 2006;Kaya & Tuncer 2007;Roberts et al 2009), using algorithms such as weight perturbation (Jabri & Flower 1992) or CMA-ES (Hansen & Ostermeier 2001). However, these methods generally have far slower convergence when compared to model-based methods, where a theoretical model is used to guide the optimization between experimental trials.…”
Section: Optimization Methodsmentioning
confidence: 99%
“…Several studies have been performed with optimization routines guiding systematic experiments or computational fluid dynamics (CFD) trials (Kern & Koumoutsakos 2006;Kaya & Tuncer 2007;Roberts et al 2009), using algorithms such as weight perturbation (Jabri & Flower 1992) or CMA-ES (Hansen & Ostermeier 2001). However, these methods generally have far slower convergence when compared to model-based methods, where a theoretical model is used to guide the optimization between experimental trials.…”
Section: Optimization Methodsmentioning
confidence: 99%
“…The most general perturbation algorithm is called serial weight perturbation [25], [26]. Weights are updated according to (1) where is the learning rate, is the perturbation magnitude, is the perturbed error, and is the nominal error.…”
Section: Learning Algorithms For Hardware Implementationsmentioning
confidence: 99%
“…For example, if 00 , , In designing a neural network (NN), we focus on the multi-layer feed-forward NN with training done off-line by smoothing the SS hysteresis function to make it differentiable and adapting the back-propagation method 20 . Other techniques such as the weight perturbation method 21 or the extreme learning machine 22 might also be suitable especially for on-chip training. The key difference with standard neural network design is the hysteresis in the threshold function which makes the overall design more robust, but requires an extra condition during the training: the currents arriving at each neuron have to be above threshold 23 .…”
Section: Fan-outmentioning
confidence: 99%