2018
DOI: 10.1016/j.conb.2017.10.006
|View full text |Cite
|
Sign up to set email alerts
|

Model-based predictions for dopamine

Abstract: Phasic dopamine responses are thought to encode a prediction-error signal consistent with model-free reinforcement learning theories. However, a number of recent findings highlight the influence of model-based computations on dopamine responses, and suggest that dopamine prediction errors reflect more dimensions of an expected outcome than scalar reward value. Here, we review a selection of these recent results and discuss the implications and complications of model-based predictions for computational theories… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

7
121
2

Year Published

2018
2018
2023
2023

Publication Types

Select...
5
3
1

Relationship

3
6

Authors

Journals

citations
Cited by 139 publications
(130 citation statements)
references
References 69 publications
7
121
2
Order By: Relevance
“…These results are consistent with the proposal that the midbrain dopamine system signals a generalized prediction error, reflecting a failure to predict features of an unexpected event beyond and even orthogonal to value (Gardner et al, 2018;Howard and Kahnt, 2018;Langdon et al, 2017;Takahashi et al, 2017). Importantly this proposal is not necessarily contrary to current canon; it can account for value errors as a special example of a more general function (Gardner et al, 2018), one readily apparent in the firing of individual neurons perhaps due to the priority given to such information when it is the goal of the experimental subject.…”
Section: Discussionsupporting
confidence: 90%
See 1 more Smart Citation
“…These results are consistent with the proposal that the midbrain dopamine system signals a generalized prediction error, reflecting a failure to predict features of an unexpected event beyond and even orthogonal to value (Gardner et al, 2018;Howard and Kahnt, 2018;Langdon et al, 2017;Takahashi et al, 2017). Importantly this proposal is not necessarily contrary to current canon; it can account for value errors as a special example of a more general function (Gardner et al, 2018), one readily apparent in the firing of individual neurons perhaps due to the priority given to such information when it is the goal of the experimental subject.…”
Section: Discussionsupporting
confidence: 90%
“…However, the same neurons also respond to errors in predicting the features of rewarding events, even when their value remains unchanged (Howard and Kahnt, 2018;Takahashi et al, 2017). Such sensory prediction errors would be useful for learning detailed information about the relationships between real-world events (Gardner et al, 2018;Howard and Kahnt, 2018;Langdon et al, 2017;Takahashi et al, 2017). Indeed, dopamine transients facilitate learning such relationships, independent of value, when they are appropriately positioned to mimic endogenous errors Keiflin et al, 2019;Sharpe et al, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…Recent work suggests that these representation-mediated phenomena are dopaminergically mediated. For example, behavioral over-expectations based on the rats' associatively learned model of reward contingencies are driven by dopamine signaling [94][95][96].…”
Section: Basic Preclinical Mechanismsmentioning
confidence: 99%
“…The striatum is hypothesized to receive reward prediction error (RPE) signals -- the difference between received and expected rewards -- from midbrain dopamine neurons (Barto, 1995; Montague et al, 1996; Schultz et al, 1997). The most common description of an RPE is as a “model-free” error, computed relative to the scalar value of a particular action, which itself reflects a common-currency based on a running average of previous rewards contingent on that action (Langdon et al, 2017). However, recent work suggests that RPE signals in the striatum can also reflect “model-based” information (Daw et al, 2011), where the prediction error is based on an internal simulation of future states.…”
Section: Introductionmentioning
confidence: 99%