2021
DOI: 10.48550/arxiv.2111.08792
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

PredProp: Bidirectional Stochastic Optimization with Precision Weighted Predictive Coding

Abstract: We present PredProp, a method for bidirectional, parallel and local optimisation of weights, activities and precision in neural networks. PredProp jointly addresses inference and learning, scales learning rates dynamically and weights gradients by the curvature of the loss function by optimizing prediction error precision. PredProp optimizes network parameters with Stochastic Gradient Descent and error forward propagation based strictly on prediction errors and variables locally available to each layer. Neighb… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…The learning rules used by PC networks (PCNs) require only local and Hebbian updates (Millidge, Tschantz, Seth & Buckley, 2020b) and a variety of neural microcircuits have been proposed that can implement the computations required by PC (Bastos et al, 2012;Keller & Mrsic-Flogel, 2018). Moreover, recent works have begun exploring the use of large-scale PCNs in machine learning tasks, to some success (Kinghorn, Millidge & Buckley, 2021;Lotter, Kreiman & Cox, 2016;Millidge, 2019;Ofner & Stober, 2021;Salvatori, Pinchetti et al, 2022;Salvatori et al, 2021). Unlike the other algorithms presented here, PC has a mathematical interpretation as in terms of variational Bayesian inference (Bogacz, 2017;Buckley, Kim, McGregor & Seth, 2017;Friston, 2003Friston, , 2005Millidge, Seth & Buckley, 2021), and the variables in the model can be mapped to explicit probabilistic elements of a generative model.…”
Section: Predictive Codingmentioning
confidence: 99%
“…The learning rules used by PC networks (PCNs) require only local and Hebbian updates (Millidge, Tschantz, Seth & Buckley, 2020b) and a variety of neural microcircuits have been proposed that can implement the computations required by PC (Bastos et al, 2012;Keller & Mrsic-Flogel, 2018). Moreover, recent works have begun exploring the use of large-scale PCNs in machine learning tasks, to some success (Kinghorn, Millidge & Buckley, 2021;Lotter, Kreiman & Cox, 2016;Millidge, 2019;Ofner & Stober, 2021;Salvatori, Pinchetti et al, 2022;Salvatori et al, 2021). Unlike the other algorithms presented here, PC has a mathematical interpretation as in terms of variational Bayesian inference (Bogacz, 2017;Buckley, Kim, McGregor & Seth, 2017;Friston, 2003Friston, , 2005Millidge, Seth & Buckley, 2021), and the variables in the model can be mapped to explicit probabilistic elements of a generative model.…”
Section: Predictive Codingmentioning
confidence: 99%