2020
DOI: 10.1037/xge0000749
|View full text |Cite
|
Sign up to set email alerts
|

It’s new, but is it good? How generalization and uncertainty guide the exploration of novel options.

Abstract: How do people decide whether to try out novel options as opposed to tried-and-tested ones? We argue that they infer a novel option's reward from contextual information learned from functional relations and take uncertainty into account when making a decision. We propose a Bayesian optimization model to describe their learning and decision making. This model relies on similarity-based learning of functional relationships between features and rewards, and a choice rule that balances exploration and exploitation … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
44
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
4
4
1

Relationship

4
5

Authors

Journals

citations
Cited by 50 publications
(45 citation statements)
references
References 136 publications
(209 reference statements)
1
44
0
Order By: Relevance
“…Instead, less efficient random exploration strategies must be used (e.g., softmax exploration). A combined model of GP regression with upper confidence sampling has proved to be an effective model in a wide number of contexts, describing how people explore different food options based on real world data (Schulz, Bhui, et al, 2019), predicting whether or not to people will try out novel options (Stojić, Schulz, Analytis, & Speekenbrink, 2020), and explaining developmental differences between how children and adults search for rewards (Meder, Wu, Schulz, & Ruggeri, 2020;Schulz, Wu, Ruggeri, & Meder, 2018).…”
Section: Using Function Learning To Guide Searchmentioning
confidence: 99%
“…Instead, less efficient random exploration strategies must be used (e.g., softmax exploration). A combined model of GP regression with upper confidence sampling has proved to be an effective model in a wide number of contexts, describing how people explore different food options based on real world data (Schulz, Bhui, et al, 2019), predicting whether or not to people will try out novel options (Stojić, Schulz, Analytis, & Speekenbrink, 2020), and explaining developmental differences between how children and adults search for rewards (Meder, Wu, Schulz, & Ruggeri, 2020;Schulz, Wu, Ruggeri, & Meder, 2018).…”
Section: Using Function Learning To Guide Searchmentioning
confidence: 99%
“…However, the hidden structure we have specified is a linear subspace lying within a non-linear (quadratic) mapping. Task 1 also shares similarities with the Feature-based Multi-Armed Bandit (FMAB) task of Stojic et al [16] in that the reward probability is a function of bivariate stimuli. However, FMAB uses a linear function and participants make a multi-way (rather than binary) decision on each trial.…”
Section: Experimental Designmentioning
confidence: 97%
“…5d, 99.9% CI: 0.53, 0.84). Thus, customers took contextual features into account to guide their exploration, similar to findings in contextual bandit tasks 23,24 .…”
Section: /15mentioning
confidence: 97%
“…Schulz et al 23 investigated how contextual information (an option's features) can aid generalization and exploration in tasks where the context is linked to an option's quality by an underlying function. Participants used a combination of functional generalization and directed exploration to learn the underlying mapping from context to reward (see also 21,24 ).…”
Section: Generalizationmentioning
confidence: 99%