Proceedings of the Genetic and Evolutionary Computation Conference 2019
DOI: 10.1145/3321707.3321723
|View full text |Cite
|
Sign up to set email alerts
|

Learning with delayed synaptic plasticity

Abstract: The plasticity property of biological neural networks allows them to perform learning and optimize their behavior by changing their configuration. Inspired by biology, plasticity can be modeled in artificial neural networks by using Hebbian learning rules, i.e. rules that update synapses based on the neuron activations and reinforcement signals. However, the distal reward problem arises when the reinforcement signals are not available immediately after each network output to associate the neuron activations th… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2

Relationship

3
2

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 30 publications
0
3
0
Order By: Relevance
“…In Orchard and Wang [37], linear and non-linear learning rules are evolved to adapt to a simple foraging task. Yaman et al [62] use genetic algorithms to optimize delayed synaptic plasticity that can learn from distal rewards. Common to these examples are that learning rules that update neural connections have access to a reward signal during the lifetime of the agent.…”
Section: Hebbian Plasticitymentioning
confidence: 99%
“…In Orchard and Wang [37], linear and non-linear learning rules are evolved to adapt to a simple foraging task. Yaman et al [62] use genetic algorithms to optimize delayed synaptic plasticity that can learn from distal rewards. Common to these examples are that learning rules that update neural connections have access to a reward signal during the lifetime of the agent.…”
Section: Hebbian Plasticitymentioning
confidence: 99%
“…The eligibility traces were proposed to trace the pairwise activations of pre-and post-synaptic neurons during an episode [3]. Data structures inspired by the eligibility traces were previously employed to associate the pairwise neuron activations with reinforcement signals [7,16,18]. Shown in Table 1, we use neuron activation traces (NATs) in each synapse to keep track of their activations (i.e.…”
Section: Evolving Plasticity For Producing Noveltymentioning
confidence: 99%
“…MCC was demonstrated for the very first time in a maze navigation problem [13]. Mazes are, in fact, a paradigmatic example of tasks with sparse, delayed reward [15], for which various approaches based on quality search [16], [17] or novelty [18] have been proposed. According to the setting proposed in [13], tasks (mazes) are co-evolved with agents (maze navigators) controlled by neural networks.…”
Section: Introductionmentioning
confidence: 99%