2016
DOI: 10.1145/2897824.2925881
|View full text |Cite
|
Sign up to set email alerts
|

Terrain-adaptive locomotion skills using deep reinforcement learning

Abstract: Reinforcement learning offers a promising methodology for developing skills for simulated characters, but typically requires working with sparse hand-crafted features. Building on recent progress in deep reinforcement learning (DeepRL), we introduce a mixture of actor-critic experts (MACE) approach that learns terrain-adaptive dynamic locomotion skills using high-dimensional state and terrain descriptions as input, and parameterized leaps or steps as output actions. MACE learns more quickly than a single actor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
230
0
2

Year Published

2017
2017
2023
2023

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 247 publications
(232 citation statements)
references
References 45 publications
0
230
0
2
Order By: Relevance
“…biped to quadruped), entirely different skeletal structure, different body proportions, way of movement and behaviour. Another way to address the design of efficient locomotion controllers, which requires no other a priori information other than the mechanical structure of the creature, is deep learning; a great effort has been devoted in this direction over the last few years [PBYvdP17, HKS17].…”
Section: Discussionmentioning
confidence: 99%
“…biped to quadruped), entirely different skeletal structure, different body proportions, way of movement and behaviour. Another way to address the design of efficient locomotion controllers, which requires no other a priori information other than the mechanical structure of the creature, is deep learning; a great effort has been devoted in this direction over the last few years [PBYvdP17, HKS17].…”
Section: Discussionmentioning
confidence: 99%
“…Finally, the current framework does not utilize any machine learning techniques. The reinforcement and other learning approaches with well‐specified reward functions can learn robust control policies from a broad range of kinematic reference motions [PALvdP18, PBYVDP17, LH17, LPY16]. However, motion capture for sports with vast moving space in the outdoor environment, such as skiing, snowboarding and waterskiing, is usually not feasible.…”
Section: Resultsmentioning
confidence: 99%
“…Kang et al [KL17] used motion capture data, but they changed the end‐effector positions to contact with the complex environment properly. Peng et al [PBYVDP17, PALvdP18] mimicked motion capture reference data using DRL.…”
Section: Related Workmentioning
confidence: 99%