2018
DOI: 10.48550/arxiv.1812.03201
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Residual Reinforcement Learning for Robot Control

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(24 citation statements)
references
References 0 publications
0
24
0
Order By: Relevance
“…A general avenue to addressing the sample complexity in RL is the deliberate use of inductive bias or prior knowledge to aid the exploratory process. This includes reward-shaping [5], [6], [7], curriculum learning [8], learning from demonstrations [9], [10], and the use of behavioural priors [11], [12], [13], [14]. The incorporation of prior knowledge in the form of behavioural priors has been gaining increasing traction in recent years.…”
Section: Learned Controllers Classical Controllersmentioning
confidence: 99%
See 3 more Smart Citations
“…A general avenue to addressing the sample complexity in RL is the deliberate use of inductive bias or prior knowledge to aid the exploratory process. This includes reward-shaping [5], [6], [7], curriculum learning [8], learning from demonstrations [9], [10], and the use of behavioural priors [11], [12], [13], [14]. The incorporation of prior knowledge in the form of behavioural priors has been gaining increasing traction in recent years.…”
Section: Learned Controllers Classical Controllersmentioning
confidence: 99%
“…RLBP approaches can directly query an action from the prior at any given state. This allows for a diverse range of mechanisms for introducing inductive bias during training, including regularisation [15], [16], [17], exploration bias [12], [13], [17] and residual learning [11], [18].…”
Section: Learned Controllers Classical Controllersmentioning
confidence: 99%
See 2 more Smart Citations
“…We are interested in applications to robotics control, which typically have continuous state and action spaces (Collins et al, 2005;Abbeel et al, 2007;Levine et al, 2016). For example, reinforcement learning can be used to learn controllers when the dynamics are unknown (Abbeel et al, 2007;Ross & Bagnell, 2012) or partially unknown (e.g., due to unknown disturbances such as wind or friction) (Akametalu et al, 2014;Berkenkamp et al, 2017;Johannink et al, 2018). Understanding sample complexity is especially important in this application, since the eventual goal is for robots to be able to learn based on real world experience, which can be very costly to obtain.…”
Section: Introductionmentioning
confidence: 99%