2020 21st Asia-Pacific Network Operations and Management Symposium (APNOMS) 2020
DOI: 10.23919/apnoms50412.2020.9237060
|View full text |Cite
|
Sign up to set email alerts
|

Deep Multi-Agent Reinforcement Learning for Resource Allocation in D2D Communication Underlaying Cellular Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
3
1

Relationship

0
7

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 10 publications
0
8
0
Order By: Relevance
“…RRAM in such networks is one major concern, especially for mmWave-based cellular networks, as the D2D links require frequent link re-establishment to combat the high blockage rate. The authors in [172] propose a multi-agent Double DQNbased scheme to address the problem of joint subcarrier assignment and power allocation in D2D underlying 5G cellular networks. The agents in their model are the D2D pairs, whose action space is discrete, corresponding to determining the transmit power allocation on the available subcarriers.…”
Section: ) In Cellular Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…RRAM in such networks is one major concern, especially for mmWave-based cellular networks, as the D2D links require frequent link re-establishment to combat the high blockage rate. The authors in [172] propose a multi-agent Double DQNbased scheme to address the problem of joint subcarrier assignment and power allocation in D2D underlying 5G cellular networks. The agents in their model are the D2D pairs, whose action space is discrete, corresponding to determining the transmit power allocation on the available subcarriers.…”
Section: ) In Cellular Networkmentioning
confidence: 99%
“…In particular, depending on the type of radio resources under investigation, resources with continuous nature such as power typically implement policy-based algorithms, while resources with discreet nature such as channel allocation and user association typically implement value-based algorithms. Simultaneous dealing with continuous and discrete types of radio resources may integrate both the policy-and value-based DRL algorithms to learn a global policy as in [41], [182], or even adopting the value-based algorithms as in, e.g., [172], [174], [185] with an expense of added quantization error.…”
Section: Findings and Lessons Learnedmentioning
confidence: 99%
“…The authors in [164] propose a multi-agent Double DQNbased scheme to address the problem of joint subcarrier assignment and power allocation in D2D underlying 5G cellular networks. The agents in their model are the D2D pairs, whose action space is discrete, corresponding to determining the transmit power allocation on the available subcarriers.…”
Section: ) In Cellular and Homnetsmentioning
confidence: 99%
“…The idea of leveraging the neighbors' information for centralized training is also considered in [81]. However, different from [80] that uses the BSs to collect information, the authors in [81] consider that each D2D user is able to collect information from neighbors, including interfering neighbors' states and interfered neighbors' states. These information can then be fed to a DDQN based agent model to learn an optimal spectrum selection policy.…”
Section: A Dynamic Spectrum Accessmentioning
confidence: 99%
“…The simulation results demonstrate that the proposed dedicated DQN is able to achieve better results than that of the shared DQN used in [89] and the random baseline in terms of sum capacity and the successful transmission probability of the V2V links. The agent information collecting mechanism utilized in the above works for centralized training, e.g., [80] and [81], do not consider a preprocess of state information, which may lead to extra signaling overhead and increase the computational complexity at the BS. The state information preprocess is investigated in [58].…”
Section: A Dynamic Spectrum Accessmentioning
confidence: 99%