2016
DOI: 10.1007/978-3-319-46227-1_35
|View full text |Cite
|
Sign up to set email alerts
|

Anti Imitation-Based Policy Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2020
2020
2020
2020

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 19 publications
0
1
0
Order By: Relevance
“…Traditional IRL methods learn from (near) optimal demonstration. More recent approaches extend IRL to learn from other types of observations, e.g., a set of (nonnecessarily optimal) demonstrations rated by an expert [El Asri et al, 2016;Burchfield et al, 2016], bad demonstrations [Sebag et al, 2016] or pairwise comparisons [da Silva et al, 2006;Wirth and Neumann, 2015]. In the latter case, the interactive setting is investigated with a reliable expert [Chernova and Veloso, 2009] or unreliable one [Weng et al, 2013].…”
Section: Reward Learningmentioning
confidence: 99%
“…Traditional IRL methods learn from (near) optimal demonstration. More recent approaches extend IRL to learn from other types of observations, e.g., a set of (nonnecessarily optimal) demonstrations rated by an expert [El Asri et al, 2016;Burchfield et al, 2016], bad demonstrations [Sebag et al, 2016] or pairwise comparisons [da Silva et al, 2006;Wirth and Neumann, 2015]. In the latter case, the interactive setting is investigated with a reliable expert [Chernova and Veloso, 2009] or unreliable one [Weng et al, 2013].…”
Section: Reward Learningmentioning
confidence: 99%