2016
DOI: 10.1109/comst.2016.2539923
|View full text |Cite
|
Sign up to set email alerts
|

A Survey on Applications of Model-Free Strategy Learning in Cognitive Wireless Networks

Abstract: Abstract-The framework of cognitive wireless radio is expected to endow the wireless devices with the cognitionintelligence ability, with which they can efficiently learn and respond to the dynamic wireless environment. In many practical scenarios, the complexity of network dynamics makes it difficult to determine the network evolution model in advance. As a result, the wireless decision-making entities may face a black-box network control problem and the model-based network management mechanisms will be no lo… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
36
0

Year Published

2017
2017
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 77 publications
(36 citation statements)
references
References 193 publications
(545 reference statements)
0
36
0
Order By: Relevance
“…In [13], a binary Spectrum Sensing feedback has been used to enable CRs apply a Reinforcement Learning procedure, the Q-Learning, to regulate the aggregated interference to the PU and in [14] a distributed channel admission solution is formulated based on Dynamic Programming, while in their previous work [15] a SU power control policy is also included using the same formulation but without elaborating on the belief factor enhancement. In [16], a methodical overview of all the Reinforcement Learning applications in CRNs based on the Markov Decision Process framework is provided. Additionally, the authors of [6] proposed a CPM based learning algorithm where probing the PU system targets to both learning interference channel matrices and maximizing the SNR at the SU receiver side in an underlay cognitive BF scenario.…”
Section: B Structurementioning
confidence: 99%
“…In [13], a binary Spectrum Sensing feedback has been used to enable CRs apply a Reinforcement Learning procedure, the Q-Learning, to regulate the aggregated interference to the PU and in [14] a distributed channel admission solution is formulated based on Dynamic Programming, while in their previous work [15] a SU power control policy is also included using the same formulation but without elaborating on the belief factor enhancement. In [16], a methodical overview of all the Reinforcement Learning applications in CRNs based on the Markov Decision Process framework is provided. Additionally, the authors of [6] proposed a CPM based learning algorithm where probing the PU system targets to both learning interference channel matrices and maximizing the SNR at the SU receiver side in an underlay cognitive BF scenario.…”
Section: B Structurementioning
confidence: 99%
“…In [15], different learning techniques, which are suitable for Internet of Things (IoT), are presented, taking into account the unique characteristics of IoT, including resource constraints and strict quality-of-service requirements, and studies on learning for IoT are also reviewed in [16]. The applications of machine learning in cognitive radio (CR) environments are investigated in [17] and [18]. Specifically, authors in [17] classify those applications into decisionmaking tasks and classification tasks, while authors in [18] mainly concentrate on model-free strategic learning.…”
mentioning
confidence: 99%
“…In practice, the optimization/game frameworks can however very complex, in which the utility function may not have a closed form expression and even may take discrete values. In such contexts, model-free strategy learning algorithms are very appealing approaches [3]. Players neither try to model the environment nor try to have a specific/explicit utility form.…”
Section: Arxiv:171101788v1 [Csgt] 6 Nov 2017mentioning
confidence: 99%
“…In model-free resource allocation schemes, developing decentralized strategies that converge to an equilibrium (if it exists), or at least finding conditions under which they converge, represent a main challenge [3]. The trial and error algorithms, proposed in [4], [12] and then applied to various resource sharing problems e.g.…”
Section: Arxiv:171101788v1 [Csgt] 6 Nov 2017mentioning
confidence: 99%
See 1 more Smart Citation