2013
DOI: 10.3389/fnbot.2013.00009
|View full text |Cite
|
Sign up to set email alerts
|

An intrinsic value system for developing multiple invariant representations with incremental slowness learning

Abstract: Curiosity Driven Modular Incremental Slow Feature Analysis (CD-MISFA;) is a recently introduced model of intrinsically-motivated invariance learning. Artificial curiosity enables the orderly formation of multiple stable sensory representations to simplify the agent's complex sensory input. We discuss computational properties of the CD-MISFA model itself as well as neurophysiological analogs fulfilling similar functional roles. CD-MISFA combines 1. unsupervised representation learning through the slowness princ… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2013
2013
2021
2021

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 67 publications
0
16
0
Order By: Relevance
“…In other words, a valuable policy maximizes the entropy over the final states and the expected utility over those states. This decomposition of the value of a policy fits neatly with accounts of intrinsic and extrinsic reward ( Luciw et al 2013 ) and connects to classic notions of exploration and exploitation ( Cohen et al 2007 ; Daw 2009 ). Here, increasing the entropy over goal states corresponds to the concept of a novelty bonus ( Kakade and Dayan 2002 ) or information gain, whereas maximizing expected utility corresponds to exploitation.…”
Section: Methodsmentioning
confidence: 68%
“…In other words, a valuable policy maximizes the entropy over the final states and the expected utility over those states. This decomposition of the value of a policy fits neatly with accounts of intrinsic and extrinsic reward ( Luciw et al 2013 ) and connects to classic notions of exploration and exploitation ( Cohen et al 2007 ; Daw 2009 ). Here, increasing the entropy over goal states corresponds to the concept of a novelty bonus ( Kakade and Dayan 2002 ) or information gain, whereas maximizing expected utility corresponds to exploitation.…”
Section: Methodsmentioning
confidence: 68%
“…An important difference is, however, that exploration is often equated with random or stochastic behavior in reinforcement learning schemes (but see Thrun, 1992), whereas in our framework, maximizing entropy over outcome states is a goal-driven, purposeful process—with the aim of accessing allowable states. Furthermore, this distinction neatly reflects the differentiation between intrinsic and extrinsic reward (Schmidhuber, 1991, 2009; Luciw et al, 2013), where extrinsic reward refers to externally administered reinforcement—corresponding to maximizing expected utility—and intrinsic reward is associated with maximizing entropy over outcomes. Maximizing intrinsic reward is usually associated with seeking new experiences in order to increase context-sensitive learning—which is reflected as increasing model-evidence or minimizing surprise in the active inference framework.…”
Section: Boredom and Novelty Seeking Under The Free Energy Principlementioning
confidence: 99%
“…33,36 When dealing with real-time video sequences, the limit or length of data is not known a priori, therefore, an incremental algorithm is needed. A various incremental version of SFA was proposed in the literature such as by Luciw et al 34,35 which learns the estimation of the features extracted resulting loss of accuracy.…”
Section: Motion Primitive Segmentationmentioning
confidence: 99%