2022
DOI: 10.1109/taes.2021.3130832
|View full text |Cite
|
Sign up to set email alerts
|

Intelligent Action Selection for NGSO Networks With Interference Constraints: A Modified Q-Learning Approach

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
8
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 11 publications
(8 citation statements)
references
References 35 publications
0
8
0
Order By: Relevance
“…For limited power resources, dynamic power allocation is more suitable for accommodating varying traffic demands than fixed power allocation in practical systems. Moreover, dynamic power allocation is also an effective way to mitigate interference, and it has been applied to diverse scenarios, such as GEO and LEO coexistence scenario 15,16 and LEO and terrestrial network coexistence scenario 17 . For the GEO and LEO coexistence scenario, the guiding principle of LEO satellite deployment is that the LEO system cannot cause unacceptable interference to the GEO system.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…For limited power resources, dynamic power allocation is more suitable for accommodating varying traffic demands than fixed power allocation in practical systems. Moreover, dynamic power allocation is also an effective way to mitigate interference, and it has been applied to diverse scenarios, such as GEO and LEO coexistence scenario 15,16 and LEO and terrestrial network coexistence scenario 17 . For the GEO and LEO coexistence scenario, the guiding principle of LEO satellite deployment is that the LEO system cannot cause unacceptable interference to the GEO system.…”
Section: Related Workmentioning
confidence: 99%
“…So, the interference caused by LEO satellites should be mitigated by certain measures to protect GEO users. However, under the constraints of GEO network interference protection, it is difficult to maximize LEO system performance 17 . To ensure the service quality of GEO satellite users and maximize the throughput for LEO users simultaneously, the adaptive beam power control is proposed, 18,19 which can improve spectrum efficiency and ensure user fairness.…”
Section: Related Workmentioning
confidence: 99%
“…It means that frequency resources, in the physical form of communication beams (comm‐beams), are used statically. However, for NGSO networks, the distribution of comm‐beams carrying different frequencies is complex and changes dynamically in real time, due to the large scale, dynamic and multinode features 7,8 . As shown in Figure 1, the frequency subbands f1,0.1emf2,0.1emf3, and f4 occur in different spatial locations in the physical form of comm‐beams in downlink case, and the ESs include in‐motion ESs, gateway stations, user terminals, and so on.…”
Section: Introductionmentioning
confidence: 99%
“…(2020) applied Q-learning to the power allocation problem in satellite-to-ground communication using LEO satellites; Zhou et al . (2021) proposed an adaptive routing strategy based on the double Q-learning for Satellite Internet of Things (S-IoT); Ren et al . (2022) proposed a modified Q-Learning approach based on an iterative greedy algorithm to upgrade the system performance of non-geostationary orbit network.…”
Section: Introductionmentioning
confidence: 99%
“…It has been implemented in various areas. Take the field of satellite communications as an example, Jiang and Zhu (2020) proposed a long-term optimal capacity allocation algorithm to optimize the long-term utility of the multi-layer satellite networks based on Q-learning; Tsuchida et al (2020) applied Q-learning to the power allocation problem in satellite-to-ground communication using LEO satellites; Zhou et al (2021) proposed an adaptive routing strategy based on the double Qlearning for Satellite Internet of Things (S-IoT); Ren et al (2022) proposed a modified Q-Learning approach based on an iterative greedy algorithm to upgrade the system performance of nongeostationary orbit network. The reward function of Q-Learning guides the agent and determines the performance of algorithm to some extent.…”
mentioning
confidence: 99%