2012 18th Asia-Pacific Conference on Communications (APCC) 2012
DOI: 10.1109/apcc.2012.6388091
|View full text |Cite
|
Sign up to set email alerts
|

A relay selection scheme using Q-learning algorithm in cooperative wireless communications

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
14
0

Year Published

2014
2014
2021
2021

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 15 publications
(14 citation statements)
references
References 8 publications
0
14
0
Order By: Relevance
“…Specifically, the forward propagation algorithms apply the carefully trained weight matrixes and bias vectors for carrying out the associated linear and [294] reduced-state SARSA cellular network dynamic channel allocation considering both mobile traffic and call handoffs. [295] on-policy SARSA CR network distributed multiagent sensing policy relying on local interactions among SU [296] on-policy SARSA MANET energy-aware reactive routing protocol for maximizing network lifetime [297] on-policy SARSA HetNet resource management for maximizing resource utilization and guaranteeing QoS [298] approximate SARSA P2P network energy harvesting aided power allocation policy for maximizing the throughput [299] Q-learning WBAN power control scheme to mitigate interference and to improve throughput [300] Q-learning OFDM system adaptive modulation and coding not relying on off-line training from PHY [301] Q-learning cooperative network efficient relay selection scheme meeting the symbol error rate requirement [302] decentralized Q-learning CR network aggregated interference control without introducing signaling overhead [303] convergent Q-learning WSN sensors' sleep scheduling scheme for minimizing the tracking error activation operations. By contrast, the backward propagation algorithms, which are widely used in the industrial field define a so-called loss function for quantifying the difference between the output produced by the training samples' and the real output.…”
Section: Deep Learning In Wireless Networkmentioning
confidence: 99%
“…Specifically, the forward propagation algorithms apply the carefully trained weight matrixes and bias vectors for carrying out the associated linear and [294] reduced-state SARSA cellular network dynamic channel allocation considering both mobile traffic and call handoffs. [295] on-policy SARSA CR network distributed multiagent sensing policy relying on local interactions among SU [296] on-policy SARSA MANET energy-aware reactive routing protocol for maximizing network lifetime [297] on-policy SARSA HetNet resource management for maximizing resource utilization and guaranteeing QoS [298] approximate SARSA P2P network energy harvesting aided power allocation policy for maximizing the throughput [299] Q-learning WBAN power control scheme to mitigate interference and to improve throughput [300] Q-learning OFDM system adaptive modulation and coding not relying on off-line training from PHY [301] Q-learning cooperative network efficient relay selection scheme meeting the symbol error rate requirement [302] decentralized Q-learning CR network aggregated interference control without introducing signaling overhead [303] convergent Q-learning WSN sensors' sleep scheduling scheme for minimizing the tracking error activation operations. By contrast, the backward propagation algorithms, which are widely used in the industrial field define a so-called loss function for quantifying the difference between the output produced by the training samples' and the real output.…”
Section: Deep Learning In Wireless Networkmentioning
confidence: 99%
“…Q-learning is also used for relay selection based on physical layer parameters in [28]. Motivated by the works in [10], [27]- [29], we propose a cross-layer relay selection scheme using Q-learning that maximizes the link layer throughput.…”
Section: Introductionmentioning
confidence: 99%
“…This paper uses a Q-learning based relay selection scheme to select an optimal relay and reduce the latency in multi-hop wireless networks. Nowadays, there are many relay selection schemes using reinforcement learning in cooperative networks [9][10][11][12][13]. These papers are focused on using reinforcement learning to improve performance through selected relays.…”
Section: Introductionmentioning
confidence: 99%