2017
DOI: 10.48550/arxiv.1707.01875
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Calibrated Fairness in Bandits

Abstract: We study fairness within the stochastic, multi-armed bandit (MAB) decision making framework. We adapt the fairness framework of "treating similar individuals similarly" [5] to this se ing. Here, an 'individual' corresponds to an arm and two arms are 'similar' if they have a similar quality distribution. First, we adopt a smoothness constraint that if two arms have a similar quality distribution then the probability of selecting each arm should be similar. In addition, we de ne the fairness regret, which corres… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

2
48
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
8
1

Relationship

1
8

Authors

Journals

citations
Cited by 54 publications
(50 citation statements)
references
References 5 publications
2
48
0
Order By: Relevance
“…Their regret bound is O( K 3 T ln T K δ ), with a higher cubic dependence on K, which is the price they pay to give equal probability to all the arms in the set. More generally, Liu et al (2017) considers a fairness through awareness. An algorithm is said to be (ǫ 1 , ǫ 2 , δ)-smooth fair 5 , if for any pair of arms a, a ′ and any t, with a probability at least…”
Section: Fairnessmentioning
confidence: 99%
See 1 more Smart Citation
“…Their regret bound is O( K 3 T ln T K δ ), with a higher cubic dependence on K, which is the price they pay to give equal probability to all the arms in the set. More generally, Liu et al (2017) considers a fairness through awareness. An algorithm is said to be (ǫ 1 , ǫ 2 , δ)-smooth fair 5 , if for any pair of arms a, a ′ and any t, with a probability at least…”
Section: Fairnessmentioning
confidence: 99%
“…Note thatLiu et al (2017) considers a general divergence function for the distribution of action selection and rewards, we introduce a special case here for a cleaner form.…”
mentioning
confidence: 99%
“…Sub-linear fairness regret and reward regret are also guaranteed in this setting. Similarly, Liu et al (2017) study the stochastic bandit with a smoothness constraint that if two arms have a similar quality distribution, the probability of selecting each arm should be similar. Patil et al (2020) and Chen et al (2020) study stochastic and contextual bandit problem with fairness requirement that there is a minimum rate that each arm must been pulled.…”
Section: Related Workmentioning
confidence: 99%
“…There have also been a handful of papers (Joseph et al 2016, Liu et al 2017, Gillen et al 2018, Patil et al 2020) that study 'fairness in bandits' in a completely different context. These works enforce a fairness criterion between arms, which is relevant in settings where a 'pull' represents some resource that is allocated to that arm, and these pulls should be distributed between arms in a fair manner.…”
Section: Related Literaturementioning
confidence: 99%