2019 IEEE 60th Annual Symposium on Foundations of Computer Science (FOCS) 2019
DOI: 10.1109/focs.2019.00017
|View full text |Cite
|
Sign up to set email alerts
|

Collaborative Learning with Limited Interaction: Tight Bounds for Distributed Exploration in Multi-armed Bandits

Abstract: Best arm identification (or, pure exploration) in multi-armed bandits is a fundamental problem in machine learning. In this paper we study the distributed version of this problem where we have multiple agents, and they want to learn the best arm collaboratively. We want to quantify the power of collaboration under limited interaction (or, communication steps), as interaction is expensive in many settings. We measure the running time of a distributed algorithm as the speedup over the best centralized algorithm … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

3
58
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 28 publications
(61 citation statements)
references
References 27 publications
3
58
0
Order By: Relevance
“…Moreover, when reducing CoPE-KB to prior CoPE with classic MAB setting (all agents are solving the same classic MAB task) [20,38], our lower and upper bounds also match the existing state-of-the-art results in [38].…”
supporting
confidence: 63%
See 3 more Smart Citations
“…Moreover, when reducing CoPE-KB to prior CoPE with classic MAB setting (all agents are solving the same classic MAB task) [20,38], our lower and upper bounds also match the existing state-of-the-art results in [38].…”
supporting
confidence: 63%
“…In such applications, it is important to develop a more general CoPE model that allows heterogeneous tasks and complex reward structures, and quantitatively investigate how task similarities impact learning acceleration. Motivated by the above facts, we propose a novel Collaborative Pure Exploration in Kernel Bandit (CoPE-KB) problem, which generalizes traditional single-task CoPE problems [20,22,38] to the multi-task setting. It also generalizes the classic MAB model to allow general (linear or nonlinear) reward structures via the powerful kernel representation.…”
mentioning
confidence: 99%
See 2 more Smart Citations
“…As a team-based and student-centred educational practice, it promotes student motivation and enhances knowledge retention via teamwork and cooperation (Sung & Hwang, 2013). While collaborative learning has been introduced and practiced in co-located settings (Barmaki et al, 2019;Huang et al, 2019;Prinsen et al, 2007;Schneider et al, 2018;Sung & Hwang, 2013), as well as distributed settings (de Freitas & Griffiths, 2007;Li et al, 2008;Schaf et al, 2009;Tao et al, 2019), measuring and evaluating collaboration still remains a challenge. Fairness of group work distribution (Ng et al, 2019), rationality of collaborative conditions (Innes & Booher, 2016) and automatism of process analytics (Rosé et al, 2008) are some of the core issues that need to be considered during collaborative learning analytics, especially in relatively large teams (Bertsimas & Gupta, 2016).…”
Section: Introductionmentioning
confidence: 99%