IEEE INFOCOM 2019 - IEEE Conference on Computer Communications 2019
DOI: 10.1109/infocom.2019.8737653
|View full text |Cite
|
Sign up to set email alerts
|

Distributed Learning and Optimal Assignment in Multiplayer Heterogeneous Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
20
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
7
2
1

Relationship

1
9

Authors

Journals

citations
Cited by 35 publications
(20 citation statements)
references
References 21 publications
0
20
0
Order By: Relevance
“…The algorithms proposed in [6] and [25] are able to achieve a near-O(log T ) expected regret using the doubling trick and O(log T ) expected regret in the one-shot scenario (when T is known). Authors in [27] provide a distributed algorithm that aims to achieve optimal network throughput without any direct communication among players. However, the algorithm has a binary signalling phase that allows players to exchange information by transmitting in specific patterns, and players are also allowed to sense such transmissions from other players.…”
Section: Related Workmentioning
confidence: 99%
“…The algorithms proposed in [6] and [25] are able to achieve a near-O(log T ) expected regret using the doubling trick and O(log T ) expected regret in the one-shot scenario (when T is known). Authors in [27] provide a distributed algorithm that aims to achieve optimal network throughput without any direct communication among players. However, the algorithm has a binary signalling phase that allows players to exchange information by transmitting in specific patterns, and players are also allowed to sense such transmissions from other players.…”
Section: Related Workmentioning
confidence: 99%
“…For instance, Game of Thrones (GoT) in [31] is a fully distributed algorithm that solves the distributed assignment problem via collision and reward feedbacks. In [32], random exploration in GoT is replaced with channel orthogonalization in order to improve the reward estimation in exploration phase. Contrary to GoT, where collisions result in zero reward, authors in [33] consider the case in which colliding users receive non-zero rewards.…”
Section: B Comparison With Related Workmentioning
confidence: 99%
“…On the other hand, the works Landgren et al [2016aLandgren et al [ ,b, 2021, Martínez-Rubio et al [2019], Dubey and Pentland [2020] restricts the communication over the network, wherein agents can only communicate with their neighbors encoded by an undirected communication graph with an unknown structure. Another branch of works Bistritz et al [2020], Kalathil et al [2014, Liu et al [2013], Tibrewal et al [2019], Rosenski et al [2016], Bistritz and Leshem [2018] investigates a competitive setting where the reward becomes zero or the reward is split when two or more agents select the same arm simultaneously. These competitive models are considered for some real-world applications wherein the communication between agents is almost impossible, and sometimes agents need to learn from the collision rather than communicating information.…”
Section: Related Workmentioning
confidence: 99%