2011 11th IEEE-RAS International Conference on Humanoid Robots 2011
DOI: 10.1109/humanoids.2011.6100908
|View full text |Cite
|
Sign up to set email alerts
|

Phase-dependent trajectory optimization for CPG-based biped walking using path integral reinforcement learning

Abstract: In this study, we introduce a phase-dependent trajectory optimization method for Central Pattern Generator (CPG)-based biped walking controllers. By exploiting the syn chronization property of the CPG controller, many legged loco motion studies have shown that the CPG-based walking controller is robust against external perturbations and works well in real environments. However, due to the nonlinear dynamic property of the coupled oscillator system composed of the CPG controller and the robot, analytically desi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
12
0

Year Published

2014
2014
2024
2024

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(12 citation statements)
references
References 16 publications
0
12
0
Order By: Relevance
“…In the path integral approach, the linearized Bellman is computed along paths starting from given initial states using sampling methods. The path integral approach has been successfully applied to learning of stochastic policies for robots with large degrees of freedom [25], [26], [27], and it is best suited for optimization around stereotyped motion trajectories. However, an additional learning is needed when a new initial state or a new goal state is given.…”
Section: B Experimental Resultsmentioning
confidence: 99%
“…In the path integral approach, the linearized Bellman is computed along paths starting from given initial states using sampling methods. The path integral approach has been successfully applied to learning of stochastic policies for robots with large degrees of freedom [25], [26], [27], and it is best suited for optimization around stereotyped motion trajectories. However, an additional learning is needed when a new initial state or a new goal state is given.…”
Section: B Experimental Resultsmentioning
confidence: 99%
“…CPG's are usually model-free as they do not take into account the system's dynamics. They smoothly produce stable rhythmic patterns, have a reduced control dimensionality, are robust against external perturbations and can be easily integrated with sensorial feedback [7,8,9]. Nevertheless, hand tuning a set of CPGs is a high-complex process due to its parameters' vast state-space (see [4,5,6,7,8]) and requires detailed knowledge about the system dynamics.…”
Section: Introductionmentioning
confidence: 99%
“…They smoothly produce stable rhythmic patterns, have a reduced control dimensionality, are robust against external perturbations and can be easily integrated with sensorial feedback [7,8,9]. Nevertheless, hand tuning a set of CPGs is a high-complex process due to its parameters' vast state-space (see [4,5,6,7,8]) and requires detailed knowledge about the system dynamics. Although this dimensionality can be minimized using techniques such as dynamical assumptions, biomechanics and reduction of the controlled DOF's, a considerable amount of data is needed to improve the tuning process [3,5,7].…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations