2018
DOI: 10.48550/arxiv.1807.06158
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Generative Adversarial Imitation from Observation

Faraz Torabi,
Garrett Warnell,
Peter Stone

Abstract: Imitation from observation (IfO) is the problem of learning directly from state-only demonstrations without having access to the demonstrator's actions. The lack of action information both distinguishes IfO from most of the literature in imitation learning, and also sets it apart as a method that may enable agents to learn from a large set of previously inapplicable resources such as internet videos. In this paper, we propose both a general framework for IfO approaches and also a new IfO approach based on gene… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
95
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
3
1

Relationship

0
10

Authors

Journals

citations
Cited by 38 publications
(97 citation statements)
references
References 20 publications
2
95
0
Order By: Relevance
“…Most directly relevant to our work are adversarial deep reinforcement learning methods which learn policies that are robust to various classes of disturbances, such as adversarial observations [40,10], rewards [8,11], direct disturbances to the system [26], or combinations thereof [21]. Nevertheless, there remains a paucity of theoretical guarantees on the generalization error, and thus sample-efficiency, of such learned policies.…”
Section: Related Workmentioning
confidence: 99%
“…Most directly relevant to our work are adversarial deep reinforcement learning methods which learn policies that are robust to various classes of disturbances, such as adversarial observations [40,10], rewards [8,11], direct disturbances to the system [26], or combinations thereof [21]. Nevertheless, there remains a paucity of theoretical guarantees on the generalization error, and thus sample-efficiency, of such learned policies.…”
Section: Related Workmentioning
confidence: 99%
“…IRL algorithms propose to infer the reward function from the expert demonstration. Among IRL algorithms, one recent branch is the Adversarial Imitation Learning (AIL) [7], [26], which trains the agent to match epert's behavior via an adversarial process. Compared with Behavioral Cloning, AIL can succeed in various challenging control tasks [7].…”
Section: A Imitation Learningmentioning
confidence: 99%
“…Instead of learning from demonstrations supplied in the first-person, third-person imitation learning [32] improves upon GAIL by recovering a domain-agnostic representation of the agent's observations. Generative adversarial imitation from observation [33] learns directly from state-only demonstrations without having access to the demonstrator's actions by recovering the state-transition cost function of the expert.…”
Section: A Imitation Learning Via Inverse Reinforcement Learningmentioning
confidence: 99%