2020
DOI: 10.1038/s41562-019-0804-2
|View full text |Cite
|
Sign up to set email alerts
|

Quantum reinforcement learning during human decision-making

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
51
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4
1

Relationship

2
8

Authors

Journals

citations
Cited by 98 publications
(51 citation statements)
references
References 93 publications
0
51
0
Order By: Relevance
“…We, however, directly modeled human decision as quantum states and decision-selection as quantum interaction, bypassing the utility-based quantum theoretic approach. Our approach is similar to the value-based quantum reinforcement learning (QRL) framework by Li et al 23 25 . Li et al in their study 25 demonstrated a QRL framework to be a better alternative for mimicking human decision-making, in which 2 QRL and 12 classical RL models were compared in modeling human decision-making among healthy and cigarette-smoking subjects while performing the Iowa Gambling Task.…”
Section: Introductionmentioning
confidence: 99%
“…We, however, directly modeled human decision as quantum states and decision-selection as quantum interaction, bypassing the utility-based quantum theoretic approach. Our approach is similar to the value-based quantum reinforcement learning (QRL) framework by Li et al 23 25 . Li et al in their study 25 demonstrated a QRL framework to be a better alternative for mimicking human decision-making, in which 2 QRL and 12 classical RL models were compared in modeling human decision-making among healthy and cigarette-smoking subjects while performing the Iowa Gambling Task.…”
Section: Introductionmentioning
confidence: 99%
“…The predefined DCC area was mainly in the ACC in this study. The ACC has been related to the function of cognitive control of emotion (Bush et al, 2000;Disner et al, 2011;Li et al, 2020;Ochsner and Gross, 2005). Individuals with a greater activation in the ACC when viewing sad faces may have a deficient inhibition ability and thus require greater cognitive effort to divert attention away from negative stimuli (Disner et al, 2011).…”
Section: Ll Open Accessmentioning
confidence: 99%
“…Therefore, an adversarial cost sequence associated with each node is assumed, and a widely used variation constraint is introduced on each cost sequence. To cope with the objective of sequential decision strategy, the problem is formulated using the reinforcement learning framework (Li et al, 2020;Littman, 2015;Sutton and Barto, 2018). Specifically, the Bayesian prior method (Smith and Paté-Cornell, 2018) is employed for the model parameters and the problem is formulated as a Bayesian adversarial multi-node bandit (MNB) model.…”
Section: Introductionmentioning
confidence: 99%