2021
DOI: 10.1609/aaai.v35i11.17156
|View full text |Cite
|
Sign up to set email alerts
|

Federated Multi-Armed Bandits

Abstract: Federated multi-armed bandits (FMAB) is a new bandit paradigm that parallels the federated learning (FL) framework in supervised learning. It is inspired by practical applications in cognitive radio and recommender systems, and enjoys features that are analogous to FL. This paper proposes a general framework of FMAB and then studies two specific federated bandit models. We first study the approximate model where the heterogeneous local models are random realizations of the global model from an unknown distribu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

1
29
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 41 publications
(30 citation statements)
references
References 15 publications
1
29
0
Order By: Relevance
“…Since most of the MAB algorithms discussed in this paper are recent [143], [144], [147], [149], it remains interesting to see their implications on practical applications, for example, quantifying the effect of bounded communication resources or energy used in wearable devices and congestion between edge nodes. Similarly, quantifying the improvement in regret bounds on actual and Quality of Experience (QoE) metrics can be promising.…”
Section: Communication-efficient Multi-agent Reinforcement Learning 1...mentioning
confidence: 99%
See 2 more Smart Citations
“…Since most of the MAB algorithms discussed in this paper are recent [143], [144], [147], [149], it remains interesting to see their implications on practical applications, for example, quantifying the effect of bounded communication resources or energy used in wearable devices and congestion between edge nodes. Similarly, quantifying the improvement in regret bounds on actual and Quality of Experience (QoE) metrics can be promising.…”
Section: Communication-efficient Multi-agent Reinforcement Learning 1...mentioning
confidence: 99%
“…Furthermore, privacy concerns sometimes limit the ability of these local servers to share data with other servers. The work in [149] studies the case of a set of servers that run a recommender system for their prospective clients. The goal of each one is to recommend the most popular content across all servers.…”
mentioning
confidence: 99%
See 1 more Smart Citation
“…Therefore, there has been much research interest in developing new FL algorithms that exhibit superior convergence (in terms of communication rounds) compared to FedAvg in the face of non-IID client data. Although most works consider training Deep [2] Averages client models x x O(|x|) FedProx [6] Proximal term for local objectives x x O(2|x|) FedMAX [7] Max-entropy term for local objectives x x O(|x|) AdaptiveFedOpt [3] Server-only optimiser x x O(|x|) MFL [8] Averages client models and optimisers x, s x, s O(|x| + |s|) Mimelite [9] Unbiased global optimiser Neural Networks (DNNs) in a round-based synchronous fashion, some works propose asynchronous algorithms to reduce training time [10], [11], and for other models such as Random Forests [12] and Multi-Armed Bandits [13].…”
Section: Introductionmentioning
confidence: 99%
“…(b) The evaluation result in scenario III. 12 ||| IEEE VEHICULAR TECHNOLOGY MAGAZINE | MONTH 2023algorithm[15], which is inspired by the principle of federated learning: training a model across multiple decentralized agents and merging the locally trained models in a communication-efficient and data-private way. Another related challenge to a federated MAB is the adversarial users who could attack the systems by sharing malicious local models, which calls for designing defense mechanisms to protect the global model.…”
mentioning
confidence: 99%