2021
DOI: 10.1101/2021.02.07.430001
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What is the state space of the world for real animals?

Abstract: A key concept in reinforcement learning (RL) is that of a state space. A state space is an abstract representation of the world using which the statistical relations in the world can be described. The simplest form of RL, model free RL, is widely applied to explain animal behavior in numerous neuroscientific studies. More complex RL versions assume that animals build and store an explicit model of the world in memory. Inherent to the application of both these classes of RL to animal behavior are assumptions ab… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(5 citation statements)
references
References 97 publications
(178 reference statements)
0
5
0
Order By: Relevance
“…This model is not designed to handle prolonged time horizons that might span multiple trials (15). Furthermore, the splitting of experience into discrete, equally-fine sub-states becomes ever more artificial as inter-trial intervals get larger and more variable (51, 52).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…This model is not designed to handle prolonged time horizons that might span multiple trials (15). Furthermore, the splitting of experience into discrete, equally-fine sub-states becomes ever more artificial as inter-trial intervals get larger and more variable (51, 52).…”
Section: Resultsmentioning
confidence: 99%
“…5A; (52)). Furthermore, the splitting of experience into discrete, equally-fine sub-states becomes ever more artificial as inter-trial intervals get longer and more variable (53, 54).…”
Section: Resultsmentioning
confidence: 99%
“…Another formalism within the general framework of RL is a semi-Markov/Markov renewal process model that explicitly learns the distribution of time intervals between consecutive state transitions (Bradtke & Duff, 1994; Daw et al, 2006; Namboodiri, 2021). Though very similar, there is one difference between a semi-Markov and Markov renewal process-based state space.…”
Section: A Discussion Of Alternative Framework and Current Limitationsmentioning
confidence: 99%
“…As the value of states is tied to the currently active state in a semi-Markov model, the value is stationary for the duration of this entire state. However, in a Markov renewal process, the value function can be defined in continuous time, as the states and transition times are treated separately (Namboodiri, 2021). While these models avoid any issues with breaking up the flow of time into states, they nevertheless suffer from some limitations.…”
Section: Figure 3 Timescale-invariance Of Behavioral Learningmentioning
confidence: 99%
“…Importantly, the ITI was similarly split to 30 states (the maximum possible ITI was divided into 3 s states starting after the previous outcome state). We have previously noted that this commonly used state space makes problematic assumptions about real animals, 85 but nevertheless use it here for its simplicity in illustrating our main claim about learning rate. The task was simulated based on the temporal parameters listed above and the timeline for the task was converted to a timeline of states, with each time step lasting 3 s. The temporal difference value function and update were then applied on this state space.…”
Section: Temporal Difference (Td) Learning Simulationsmentioning
confidence: 99%