2006
DOI: 10.1177/0278364906063822
|View full text |Cite
|
Sign up to set email alerts
|

Fast Biped Walking with a Sensor-driven Neuronal Controller and Real-time Online Learning

Abstract: In this paper, we present our design and experiments on a planar biped robot under the control of a pure sensor-driven controller. This design has some special mechanical features, for example small curved feet allowing rolling action and a properly positioned center of mass, that facilitate fast walking through exploitation of the robot's natural dynamics. Our sensor-driven controller is built with biologically inspired sensor-and motor-neuron models, and does not employ any kind of position or trajectory tra… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

4
124
0

Year Published

2006
2006
2016
2016

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 130 publications
(128 citation statements)
references
References 27 publications
4
124
0
Order By: Relevance
“…Hence, they are not easily applicable to real robots. To date, various automatic gait optimization methods have been used in locomotion to design gaits, including gradient descent methods [18,3], evolutionary algorithms [1,19], particle swarm optimization [20] and many others [21,22,2,17]. We now discuss the approaches that use surrogate models to optimize robot locomotion.…”
Section: Optimization Methods In Roboticsmentioning
confidence: 99%
“…Hence, they are not easily applicable to real robots. To date, various automatic gait optimization methods have been used in locomotion to design gaits, including gradient descent methods [18,3], evolutionary algorithms [1,19], particle swarm optimization [20] and many others [21,22,2,17]. We now discuss the approaches that use surrogate models to optimize robot locomotion.…”
Section: Optimization Methods In Roboticsmentioning
confidence: 99%
“…Due to the simplicity of this approach, such methods have been successfully applied to robotics in numerous applications [6], [9], [11], [14]. However, the straightforward application to robotics is not without peril as the generation of the ∆θ j requires proper knowledge on the system, as badly chosen ∆θ j can destabilize the policy so that the system becomes instable and the gradient estimation process is prone to fail.…”
Section: A General Approaches To Policy Gradient Estimationmentioning
confidence: 99%
“…Policy gradient methods are a notable exception to this statement. Starting with the pioneering work of Gullapali, Franklin and Benbrahim [1], [2] in the early 1990s, these methods have been applied to a variety of robot learning problems ranging from simple control tasks (e.g., balancing a ball-on a beam [3], and pole-balancing [4]) to complex learning tasks involving many degrees of freedom such as learning of complex motor skills [2], [5], [6] and locomotion [7]- [14] 1 . The advantages of policy gradient methods for robotics are numerous.…”
Section: Introductionmentioning
confidence: 99%
“…The computation of the policy update is the key step here and a variety of updates have been proposed ranging from pairwise comparisons [Strens andMoore, 2001, Ng et al, 2004a] over gradient estimation using finite policy differences [Geng et al, 2006, Mitsunaga et al, 2005, Sato et al, 2002, Tedrake et al, 2005, and general stochastic optimization methods (such as Nelder-Mead [Bagnell and Schneider, 2001], cross entropy [Rubinstein and Kroese, 2004] and population-based methods [Goldberg, 1989]) to approaches coming from optimal control such as differential dynamic programming (DDP) [Atkeson, 1998] and multiple shooting approaches [Betts, 2001] as well as core reinforcement learning methods.…”
Section: Policy Searchmentioning
confidence: 99%
“…For rhythmic behaviors half-elliptical locuses have been used as a representation of the gait pattern of a robot dog [Kohl and Stone, 2004]. Neural Networks: Instead of analytically describing rhythmic movements, neural networks can be used as oscillators to learn gaits of a a two legged robot [Geng et al, 2006, Endo et al, 2008. Also a peg-in-hole (see Figure 2.1b) and a ball-balancing task as well as a navigation task [Hailu and Sommer, 1998] have been learned with neural networks as policy function approximators.…”
Section: Pre-structured Policiesmentioning
confidence: 99%