Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems 2016
DOI: 10.1145/2858036.2858425
|View full text |Cite
|
Sign up to set email alerts
|

Interface Design Optimization as a Multi-Armed Bandit Problem

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
25
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 39 publications
(25 citation statements)
references
References 27 publications
0
25
0
Order By: Relevance
“…We implemented Sarsa, which is a standard algorithm to learn how to act in many different environment state, i.e., for each given parameter configuration [97]. It differs from multi-armed bandits, which learns how to act in one unique environment state [68]. Importantly, as evoked in Section 1, Sarsa was designed to learn one optimal behaviour in relation to the goal of a task.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…We implemented Sarsa, which is a standard algorithm to learn how to act in many different environment state, i.e., for each given parameter configuration [97]. It differs from multi-armed bandits, which learns how to act in one unique environment state [68]. Importantly, as evoked in Section 1, Sarsa was designed to learn one optimal behaviour in relation to the goal of a task.…”
Section: Methodsmentioning
confidence: 99%
“…In this sense, interactive reinforcement learning relies on small, user-specific data sets, which contrasts with the large, crowdsourced data sets used in creative applications in semantic editing [25,62,107]. Lastly, interactive approaches to reinforcement learning focuses on exploring agent actions based on human feedback on actions, which contrasts with the focus on optimising one parametric state based on user feedback over states-as used in Bayesian Optimisation [13,67] or multi-armed bandits [68].…”
Section: Interactive Reinforcement Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…Online interface refinement, the category in which this paper falls, describes methods which actively change the interface based on some objective during or between interactions. This approach is readily applied in games where an optimal performance or engagement level might be achieved through game feature refinement [7,14,15]. Similarly, BIGnav [12] probabilistically fused inputs and prior information about locations on a map to improve navigation performance.…”
Section: Related Workmentioning
confidence: 99%
“…Lan and Baraniuk [8] used sparse factor analysis with bandits to identify sequences of educational content that could maximize students performance on subsequent assessments. Lomas et al [9] showed how bandits can be used to search a large space of design decisions in creating educational games. Williams et al [18] used Thompson Sampling to identify highly rated explanations for how to solve Math problems, and chose priors that assumed that every explanation was equally rated.…”
Section: Related Workmentioning
confidence: 99%