2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) 2018
DOI: 10.1109/iros.2018.8593722
|View full text |Cite
|
Sign up to set email alerts
|

Feedback Control For Cassie With Deep Reinforcement Learning

Abstract: Bipedal locomotion skills are challenging to develop. Control strategies often use local linearization of the dynamics in conjunction with reduced-order abstractions to yield tractable solutions. In these model-based control strategies, the controller is often not fully aware of many details, including torque limits, joint limits, and other non-linearities that are necessarily excluded from the control computations for simplicity. Deep reinforcement learning (DRL) offers a promising model-free approach for con… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
136
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 170 publications
(136 citation statements)
references
References 13 publications
0
136
0
Order By: Relevance
“…Deep RL has been used extensively to learn locomotion policies in simulation [4,20,34,50] and even transfer them to real-world robots [24,44], but this inevitably incurs a loss of performance due to discrepancies in the simulation, and requires accurate system identification. Using such algorithms directly in the real world has proven challenging.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Deep RL has been used extensively to learn locomotion policies in simulation [4,20,34,50] and even transfer them to real-world robots [24,44], but this inevitably incurs a loss of performance due to discrepancies in the simulation, and requires accurate system identification. Using such algorithms directly in the real world has proven challenging.…”
Section: Related Workmentioning
confidence: 99%
“…The policy is trained only on a flat terrain, but the learned gait is robust and can handle obstacles that were not seen during training. learning of locomotion gaits in simulation [4,20,34,50], requiring accurate system identification and modeling.…”
Section: Introductionmentioning
confidence: 99%
“…Schulman et al [22] trained a locomotion policy for a similar 2D walker with an actor-critic method. More recent work obtained full 3D locomotion policies [23][24][25][26]. In these papers, animated characters achieve remarkable motor skills in simulation.…”
Section: Introductionmentioning
confidence: 99%
“…A variety of approaches, starting from inverted pendulum models [2], zero moment points [3], passivity based control [4], [5], capture points [6], hybrid zero dynamics [7] to even machine learning based approaches [8], [9] have been explored. A more recent trend has been the use of deep reinforcement learning (D-RL) methods [10], [11] to determine optimal policies for efficient and robust walking behaviors in both bipeds and quadrupeds. D-RL was successfully used to achieve walking in the quadrupedal robot Minitaur [10], where the control policy was trained via one of the well known policy gradient based approaches called the proximal policy optimization (PPO) [12].…”
Section: Introductionmentioning
confidence: 99%