The 2020 Conference on Artificial Life 2020
DOI: 10.1162/isal_a_00299
|View full text |Cite
|
Sign up to set email alerts
|

Sample and time efficient policy learning with CMA-ES and Bayesian Optimisation

Abstract: In evolutionary robot systems where morphologies and controllers of real robots are simultaneously evolved, it is clear that there is likely to be requirements to refine the inherited controller of a 'newborn' robot in order to better align it to its newly generated morphology. This can be accomplished via a learning mechanism applied to each individual robot: for practical reasons, such a mechanism should be both sample and time-efficient. In this paper, We investigate two ways to improve the sample and time … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
30
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
2

Relationship

6
2

Authors

Journals

citations
Cited by 21 publications
(30 citation statements)
references
References 21 publications
0
30
0
Order By: Relevance
“…First, a novelty search algorithm is used to evolve a large list of robots, from which a subset of robots is selected to form a bootstrap population. Secondly, a learning algorithm (NIP-ES, [26]) is applied to the bootstrap population to learn a controller for these robots, to gain insight into the potential capabilities of the bootstrap population. Finally, the bootstrap population (of body plans) is used to form the initial population of an evolutionary algorithm by pairing each body plan with a random CPPN that generates a controller.…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…First, a novelty search algorithm is used to evolve a large list of robots, from which a subset of robots is selected to form a bootstrap population. Secondly, a learning algorithm (NIP-ES, [26]) is applied to the bootstrap population to learn a controller for these robots, to gain insight into the potential capabilities of the bootstrap population. Finally, the bootstrap population (of body plans) is used to form the initial population of an evolutionary algorithm by pairing each body plan with a random CPPN that generates a controller.…”
Section: Discussionmentioning
confidence: 99%
“…In these experiments, the Novelty-driven Increasing Population Evolutionary Strategy (NIP-ES) algorithm is used to train a controller for each robot in the bootstrap population, based on our previous work [26]. The NIP-ES algorithm is a modified version of the increasing-population Covariance Matrix Adaptation Evolutionary Strategy (IPOP-CMAES) [27].…”
Section: Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…10. These body plans are being used to test and develop some of the hardware, such as the Robot Fabricator [4], and software, such as selecting and developing the controller architecture and learning algorithms which will eventually be applied to evolved robots [16]. They are therefore designed to require different controllers, with a maze solving task in mind.…”
Section: Example Body Plansmentioning
confidence: 99%
“…Given that the search-space of body plans (morphologies) and controllers is very large, and printing/assembly is both time-consuming and expensive, we intend to bootstrap our evolutionary process by starting with a diverse population of manufacturable robots. For instance, it was shown by Le Goff et al [6] that for only learning it takes at least a couple hundred of evaluations to produce controllers for ARE-robots to solve tasks.…”
Section: Introductionmentioning
confidence: 99%