2021
DOI: 10.48550/arxiv.2106.09611
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

A Reinforcement Learning Approach for an IRS-assisted NOMA Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…Any negative reward will work as the agent will try to avoid such action in the future. We will use the sum of the rate deficit across all users as the negative reward [18]. The set ε contains users j = 1, .., J, ε ∈ N whose QoS are not satisfied at time-step t. Thus, we define the sum of the rate deficit across all users as…”
Section: Robust Design Problem As Td3 Environmentmentioning
confidence: 99%
See 1 more Smart Citation
“…Any negative reward will work as the agent will try to avoid such action in the future. We will use the sum of the rate deficit across all users as the negative reward [18]. The set ε contains users j = 1, .., J, ε ∈ N whose QoS are not satisfied at time-step t. Thus, we define the sum of the rate deficit across all users as…”
Section: Robust Design Problem As Td3 Environmentmentioning
confidence: 99%
“…In [17], Meng et al proposed a DRL-based solution for sum-rate maximization in multi-cell networks. Xiao et al proposed a deep deterministic policy gradient (DDPG) based solution to jointly optimize the beamforming and phase shifts of IRS elements for sum-rate maximization in an IRS-assisted MISO-NOMA system [18]. In [19], Gao et al proposed a deep Q-network (DQN) based algorithm to jointly optimize IRS phase shifts and cluster power allocation in a NOMA system using the zero forcing approach.…”
Section: Introductionmentioning
confidence: 99%
“…The work in [22] adopted the zero-forcing beamforming (ZFBF) technique while utilizing a deep Q-network (DQN) agent for optimizing phase shifts of the IRS elements. Xie et al used DDPG to jointly optimize the beamforming vectors and IRS phase shifts for the sum-rate maximization problem [23]. The work in [24] proposed a multi-agent DRL-based design that jointly optimizes the subcarrier assignment, power allocation, and IRS phase shifts in NOMA-assisted semi-grant-free systems, while the resource allocation problem for NOMAunmanned aerial vehicle system was considered in [25].…”
Section: Introductionmentioning
confidence: 99%