2018
DOI: 10.1007/978-3-319-94042-7_7
|View full text |Cite
|
Sign up to set email alerts
|

Learning to Run Challenge Solutions: Adapting Reinforcement Learning Methods for Neuromusculoskeletal Environments

Abstract: In the NIPS 2017 Learning to Run challenge, participants were tasked with building a controller for a musculoskeletal model to make it run as fast as possible through an obstacle course. Top participants were invited to describe their algorithms. In this work, we present eight solutions that used deep reinforcement learning approaches, based on algorithms such as Deep Deterministic Policy Gradient, Proximal Policy Optimization, and Trust Region Policy Optimization. Many solutions use similar relaxations and he… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
52
0

Year Published

2018
2018
2021
2021

Publication Types

Select...
5
2
1

Relationship

7
1

Authors

Journals

citations
Cited by 61 publications
(52 citation statements)
references
References 24 publications
0
52
0
Order By: Relevance
“…Various RL techniques have been effectively used since the first competition [124,125], including frame skipping, discretization of the action space, and reward shaping. These are practical techniques that constrain the problem in certain ways to encourage an agent to search successful regions faster in the initial stages of training.…”
Section: Top Solutions and Resultsmentioning
confidence: 99%
“…Various RL techniques have been effectively used since the first competition [124,125], including frame skipping, discretization of the action space, and reward shaping. These are practical techniques that constrain the problem in certain ways to encourage an agent to search successful regions faster in the initial stages of training.…”
Section: Top Solutions and Resultsmentioning
confidence: 99%
“…In the early stage of training, participants reduced accuracy to speed up simulations to train their models more quickly. Later, they fine-tuned the model by switching the accuracy to the same one used for the competition [22].…”
Section: Solutionsmentioning
confidence: 99%
“…D EEP neural networks have pushed further the envelope of reinforcement learning in a wide variety of domains, such as Atari games [1], continuous systems control [2], musculoskeletal models control for medical applications [3], etc. Deep reinforcement learning (Deep-RL) methods perform trail-and-error training through frequent interactions with the environments.…”
Section: Introductionmentioning
confidence: 99%
“…We evaluate our proposed method on a realistic physiologically-based model control task, namely Learning to Run [3]. Experimental results show that AE-DDPG outperforms not only the vanilla DDPG but also other popular RL methods in training efficiency and the resulting final policies.…”
Section: Introductionmentioning
confidence: 99%