2017
DOI: 10.1201/9781439821091
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning and Dynamic Programming Using Function Approximators

Abstract: Control systems are making a tremendous impact on our society. Though invisible to most users, they are essential for the operation of nearly all devices -from basic home appliances to aircraft and nuclear power plants. Apart from technical systems, the principles of control are routinely applied and exploited in a variety of disciplines such as economics, medicine, social sciences, and artificial intelligence.A common denominator in the diverse applications of control is the need to influence or modify the be… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
421
0
25

Year Published

2017
2017
2018
2018

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 554 publications
(447 citation statements)
references
References 121 publications
(288 reference statements)
1
421
0
25
Order By: Relevance
“…Encouraged by these successes, the Deep RL field of research is revisiting some of the main works achieved in the past three decades using deep learning function approximators (see Busoniu et al (2010) for a global view of the use of function approximators in RL). In particular, the deep learning remastering of Double Q Learning (see Hasselt et al (2015) and Hasselt (2010) for the original version without deep learning) or Memory Replay Schaul et al (2015) ideas shows that RL promises that were originally proposed decades ago are definitely worth revisiting in the light of Deep Learning architecture.…”
Section: Deep Rlmentioning
confidence: 99%
See 3 more Smart Citations
“…Encouraged by these successes, the Deep RL field of research is revisiting some of the main works achieved in the past three decades using deep learning function approximators (see Busoniu et al (2010) for a global view of the use of function approximators in RL). In particular, the deep learning remastering of Double Q Learning (see Hasselt et al (2015) and Hasselt (2010) for the original version without deep learning) or Memory Replay Schaul et al (2015) ideas shows that RL promises that were originally proposed decades ago are definitely worth revisiting in the light of Deep Learning architecture.…”
Section: Deep Rlmentioning
confidence: 99%
“…After each moment of interaction, the agent receives a feedback signal, reinforcement signal, or reward from the system being delivered to the learning system in response to the execution of control action Sutton and Barto, 1998;Busoniu et al, 2010. The most commonly studied objective is to maximize, for each time step, the expected sum of future reinforcements or discounted return defined as the sum of rewards over future time steps Sutton and Barto, 1998;Busoniu et al, 2010. A likely framework for application of the RL in power system decision and control is illustrated in Fig. 3.…”
Section: A Framework For Rl Consideration In Power System Decision Anmentioning
confidence: 99%
See 2 more Smart Citations
“…It can also be simply combined with fuzzy logic and provide the relationship between the states and the accessible action, which is the same as creating the fuzzy logic "if. ..then" engine [23][24][25].…”
Section: Introductionmentioning
confidence: 99%