2020
DOI: 10.1109/access.2020.3040748
|View full text |Cite
|
Sign up to set email alerts
|

Hierarchical Reinforcement-Learning for Real-Time Scheduling of Agile Satellites

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 13 publications
(8 citation statements)
references
References 28 publications
0
8
0
Order By: Relevance
“…Hierarchical reinforcement learning uses the decomposition of objectives into sub-objectives, which can effectively reduce the solution space and is an effective method for solving large-scale constellation decision-making problems [ 30 ]. Ren et al proposed a hierarchical reinforcement-learning algorithm based on Q-learning for response speed and stability problems in randomly occurring urgent tasks [ 31 ]. Zhao et al proposed a two-stage neural network combinatorial optimization method based on DDPG to solve the problem of temporal task assignments in dynamic environments [ 32 ].…”
Section: Introductionmentioning
confidence: 99%
“…Hierarchical reinforcement learning uses the decomposition of objectives into sub-objectives, which can effectively reduce the solution space and is an effective method for solving large-scale constellation decision-making problems [ 30 ]. Ren et al proposed a hierarchical reinforcement-learning algorithm based on Q-learning for response speed and stability problems in randomly occurring urgent tasks [ 31 ]. Zhao et al proposed a two-stage neural network combinatorial optimization method based on DDPG to solve the problem of temporal task assignments in dynamic environments [ 32 ].…”
Section: Introductionmentioning
confidence: 99%
“…However, this method is also based on supervised learning data and has limited applicability as well as problem applicability as it is also a kind of supervised learning. Ren [12] proposed a hierarchical reinforcement learning method to improve the response speed and stability of autonomous satellites to urgent tasks. The base layer is used to learn and train the network, while the top layer uses the network to assign tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Several hierarchical reinforcement approaches have been proposed to improve the performance of multi-UAV wireless networks [11]- [13]. In [11], to resolve the problem of limited data collection coverage of the backscatter sensor nodes, the hierarchical deep reinforcement learning (DRL) framework was proposed to extend the data collection coverage and minimize the total flight time of the rechargeable UAVs when performing data collecting missions.…”
Section: Introductionmentioning
confidence: 99%
“…The proposed h-DQN shows faster convergence, higher performance, and higher channel utilization than Q-learning for dynamic sensing (QADS) [14] or deep reinforcement learning for dynamic access (DRLDA) [15]. Additional, in [13], a hierarchical scheduling architecture with top-layer scheduling for satellite selection and foundation-layer precise scheduling for urgent tasks was introduced to solve the real-time earth observation satellite (EOS) scheduling problem. Here, Q-learning with an adaptive action selection strategy was proposed to solve the Markov decision process model more efficiently.…”
Section: Introductionmentioning
confidence: 99%