2018
DOI: 10.1561/9781680835397
|View full text |Cite
|
Sign up to set email alerts
|

An Introduction to Deep Reinforcement Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
273
0
1

Year Published

2020
2020
2022
2022

Publication Types

Select...
6
2

Relationship

0
8

Authors

Journals

citations
Cited by 393 publications
(274 citation statements)
references
References 0 publications
0
273
0
1
Order By: Relevance
“…Hence, a policy is commonly represented by a function approximator to overcome this difficulty [60]. The combination of RL and deep leaning (called deep RL [62]) has been successful in handling large-scale complicated tasks by using deep neural networks (DNNs) as function approximators [63], [64], but it is at the expense of complexity. Therefore, this study still considers a RL method and uses a simple neural network (NN) with one fully connected hidden layer to represent the policy (called policy network), as shown in Fig.…”
Section: Reinforcement Learning Based On-mtpmentioning
confidence: 99%
See 1 more Smart Citation
“…Hence, a policy is commonly represented by a function approximator to overcome this difficulty [60]. The combination of RL and deep leaning (called deep RL [62]) has been successful in handling large-scale complicated tasks by using deep neural networks (DNNs) as function approximators [63], [64], but it is at the expense of complexity. Therefore, this study still considers a RL method and uses a simple neural network (NN) with one fully connected hidden layer to represent the policy (called policy network), as shown in Fig.…”
Section: Reinforcement Learning Based On-mtpmentioning
confidence: 99%
“…The first layer (called input layer) is given the input values, where all the images in the state space are collected as input to this NN. The values of the middle layer (called hidden layer) are a transformation of the input values by a non-linear parametric function [62]. The last layer (called output layer) provides the output values transformed from the hidden layer, which can output an action deciding to accept or reject a new TR.…”
Section: Reinforcement Learning Based On-mtpmentioning
confidence: 99%
“…Deep reinforcement learning describes a class of goalorientated machine learning algorithms taking advantage of powerful function approximators in the context of deep learning [21,22]. Unlike supervised or unsupervised machine learning, these algorithms do not require a dedicated set of training data, since they are designed to learn from experience by interacting with their environment.…”
Section: A Deep Reinforcement Learningmentioning
confidence: 99%
“…A contour that was valid for the initial guess of the function would then have to be re-adjusted. The goal of this study is to provide a proof of principle that a deep reinforcement learning (DRL) agent (see e. g. [21] for a recent review and [22] for a standard text book on the subject) can be trained to conduct the contour deformations as needed. Such an agent could then be used in an iterative setting by deducing the contour deformation from observing the integration plane before each iteration step is conducted.…”
Section: Introductionmentioning
confidence: 99%
“…The purpose of the present study is to explore how the brain's RL machinery might utilise these opposing properties to achieve complex behavior. Much of the recent success of RL has been due to the combination of classical RL approaches with the function approximation properties of Deep Neural Networks (DNNs), known as deep RL (François-lavet et al 2018). Typically in deep RL, the action-value function Q(s, a) is represented using a DNN that takes the state s t as input and outputs the corresponding action values for that state.…”
Section: Introductionmentioning
confidence: 99%