2021
DOI: 10.1016/j.cobeha.2020.10.010
|View full text |Cite
|
Sign up to set email alerts
|

What is dopamine doing in model-based reinforcement learning?

Abstract: Experiments have implicated dopamine in model-based reinforcement learning (RL). These findings are unexpected as dopamine is thought to encode a reward prediction error (RPE), which is the key teaching signal in model-free RL. Here we examine two possible accounts for dopamine's involvement in model-based RL: the first that dopamine neurons carry a prediction error used to update a type of predictive state representation called a successor representation, the second that two well established aspects of dopami… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

1
9
0

Year Published

2021
2021
2025
2025

Publication Types

Select...
8
2

Relationship

1
9

Authors

Journals

citations
Cited by 16 publications
(10 citation statements)
references
References 100 publications
1
9
0
Order By: Relevance
“…In particular, Takahashi et al (2017) showed that dopamine carries information about the sensory features of expected rewards in rats. This evidence is consistent with a view on which SWR-based re- and preplay events have a role in both model-free and model-based value computation (e.g., Akam & Walton, 2021). Indeed, Mattar and Daw (2018) proposed that the various functional perspectives on hippocampal re- and preplay could be unified by a view that assumes that both phenomena serve the computation of long-run action values.…”
Section: Functional Perspectives On the Hedonic Representation Of Val...supporting
confidence: 89%
“…In particular, Takahashi et al (2017) showed that dopamine carries information about the sensory features of expected rewards in rats. This evidence is consistent with a view on which SWR-based re- and preplay events have a role in both model-free and model-based value computation (e.g., Akam & Walton, 2021). Indeed, Mattar and Daw (2018) proposed that the various functional perspectives on hippocampal re- and preplay could be unified by a view that assumes that both phenomena serve the computation of long-run action values.…”
Section: Functional Perspectives On the Hedonic Representation Of Val...supporting
confidence: 89%
“…To achieve this, a subset of dopamine cells are required to undergo computational processes that are more elaborate than traditional models of dopamine function predict and may in part reflect dopamine's role in sensory prediction error 12,33 . Future studies targeting the nature of any underlying circuitry encoding these detailed reinforcement signals via dopamine should include the extent to which our findings relate to proposed roles in model-based encoding 12 , or signaling of surprise from afferent sensory systems 58 that are relayed to higher-order cortical sites 32 . Alternatively, given our approach intersects appetition and aversion, it is possible that we are engaging a heterogeneous population of dopamine cells that include those that respond to high intensity sensory stimuli and are aversive when stimulated 59 .…”
Section: Discussionmentioning
confidence: 99%
“…This suggestion is in line with the idea of a stronger utilization of more explicit (model‐based) expectations in the present study, masking expectations which are based on trial‐by‐trial value associations as derived by model‐free Q learning (Botvinick & Weinstein, 2014; Daw et al, 2011; Wang, Kurth‐Nelson, et al, 2018). Besides the task demands, it might also be that assessing self‐reported expectations within each trial further contributed to participants' overall use of more explicit predictions (Akam & Walton, 2021; Doll et al, 2009; Smittenaar et al, 2013). Overall, since negative PEs were not indicative of the need for behavioral adjustment, it may be that FMθ was less affected by trial‐by‐trial predictions but more by implicit learning about the reversals, decoupling the often‐replicated association between PEs and FMθ (Cooper et al, 2019; Kaufman et al, 2010; Pinner & Cavanagh, 2017; Reber, 2013).…”
Section: Discussionmentioning
confidence: 99%