2022
DOI: 10.1088/1741-2552/ac6ca7
|View full text |Cite
|
Sign up to set email alerts
|

Emergence of associative learning in a neuromorphic inference network

Abstract: Objective. In the theoretical framework of predictive coding and active inference, the brain can be viewed as instantiating a rich generative model of the world that predicts incoming sensory data while continuously updating its parameters via minimization of prediction errors. While this theory has been successfully applied to cognitive processes – by modelling the activity of functional neural networks at a mesoscopic scale – the validity of the approach when modelling neurons as an ensemble of inferring age… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
7
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6

Relationship

2
4

Authors

Journals

citations
Cited by 12 publications
(7 citation statements)
references
References 58 publications
0
7
0
Order By: Relevance
“…The ability of LTP and LTD to reorganize granular layer activity bears a series of functional and theoretical consequences. LTP and LTD can finetune time- and frequency-dependent properties of network transmission rather than just regulating synaptic weights, bearing relevant implications for applicative scenarios [ 73 , 74 ]. Our observations lend support to the hypothesis that the granular layer processes spike bursts as a relevant computational unit, in which the composition in terms of the spike delay, the number or the frequency can be finetuned.…”
Section: Discussionmentioning
confidence: 99%
“…The ability of LTP and LTD to reorganize granular layer activity bears a series of functional and theoretical consequences. LTP and LTD can finetune time- and frequency-dependent properties of network transmission rather than just regulating synaptic weights, bearing relevant implications for applicative scenarios [ 73 , 74 ]. Our observations lend support to the hypothesis that the granular layer processes spike bursts as a relevant computational unit, in which the composition in terms of the spike delay, the number or the frequency can be finetuned.…”
Section: Discussionmentioning
confidence: 99%
“…The replication of the cerebellum’s connectivity patterns using a low-power microcontroller, 51 confirms that the evolution of neural networks drives the tuning of synapses to reduce systemic free energy. From the perspective of energy consumption, the characteristics of the brain’s energy-saving operation mechanism under the immense pressure of information processing, including optimal input and noise, excitation/inhibition balance, the size of neurons and neuron clusters, have been analyzed.…”
Section: Discussionmentioning
confidence: 70%
“…This proof of principle is a first milestone in the development of advanced neuronal networks with performance compatible with brain circuits, given their capability to compute sparse and temporally uncorrelated information. Furthermore, differently from conventional hardware, neuromorphic electronic circuits [41] can be designed to operate with a limited power consumption in multiple time domains according to circuit architectures. These advantages, deriving from an electronic implementation of biologically plausible SNNs (e.g.…”
Section: Discussionmentioning
confidence: 99%