2023
DOI: 10.1109/lcomm.2023.3266931
|View full text |Cite
|
Sign up to set email alerts
|

DDPG-Based Joint Resource Management for Latency Minimization in NOMA-MEC Networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2023
2023
2025
2025

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 17 publications
(2 citation statements)
references
References 12 publications
0
2
0
Order By: Relevance
“…This algorithm combines the representational capabilities of deep learning with the decision optimization techniques of policy gradient methods, adopting a variant of the Actor-Critic architecture, implemented through deep neural networks to approximate both policy and value functions [29]. The actor network directly maps states to deterministic actions, while the Critic network evaluates the expected return for given states and actions [30]. Additionally, DDPG incorporates an experience replay mechanism, storing past transitions (states, actions, rewards, and new states) for reuse during training, thus reducing correlations between samples and enhancing learning stability.…”
Section: Buy Tmentioning
confidence: 99%
“…This algorithm combines the representational capabilities of deep learning with the decision optimization techniques of policy gradient methods, adopting a variant of the Actor-Critic architecture, implemented through deep neural networks to approximate both policy and value functions [29]. The actor network directly maps states to deterministic actions, while the Critic network evaluates the expected return for given states and actions [30]. Additionally, DDPG incorporates an experience replay mechanism, storing past transitions (states, actions, rewards, and new states) for reuse during training, thus reducing correlations between samples and enhancing learning stability.…”
Section: Buy Tmentioning
confidence: 99%
“…The Deep Deterministic Policy Gradient (DDPG) algorithm [ 29 ], where the resource allocation problem is solved via the DDPG method.…”
Section: Simulation Analysismentioning
confidence: 99%