2017
DOI: 10.48550/arxiv.1704.04470
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Lean From Thy Neighbor: Stochastic & Adversarial Bandits in a Network

Abstract: An individual's decisions are often guided by those of his or her peers, i.e., neighbors in a social network. Presumably, being privy to the experiences of others aids in learning and decision making, but how much advantage does an individual gain by observing her neighbors? Such problems make appearances in sociology and economics and, in this paper, we present a novel model to capture such decision-making processes and appeal to the classical multi-armed bandit framework to analyze it. Each individual, in ad… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 6 publications
0
1
0
Order By: Relevance
“…Works by Bastani, Bayati, and Khosravi [2]; Kannan, Morgenstern, Roth, Waggoner, and Wu [9]; and Raghavan, Slivkins, Vaughan, and Wu [14] show that if the data is sufficiently diverse, e.g., if the contexts are randomly perturbed, then exploration may not be necessary. Celis and Salehi [7] consider a model in both the stochastic and the adversarial setting where each agent in the network plays a certain zero-regret algorithm (UCB in the stochastic setting and EXP3 in the adversarial setting) and study how much information an agent can gather from his neighbors.…”
Section: Introductionmentioning
confidence: 99%
“…Works by Bastani, Bayati, and Khosravi [2]; Kannan, Morgenstern, Roth, Waggoner, and Wu [9]; and Raghavan, Slivkins, Vaughan, and Wu [14] show that if the data is sufficiently diverse, e.g., if the contexts are randomly perturbed, then exploration may not be necessary. Celis and Salehi [7] consider a model in both the stochastic and the adversarial setting where each agent in the network plays a certain zero-regret algorithm (UCB in the stochastic setting and EXP3 in the adversarial setting) and study how much information an agent can gather from his neighbors.…”
Section: Introductionmentioning
confidence: 99%