2021
DOI: 10.48550/arxiv.2111.02371
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

What Robot do I Need? Fast Co-Adaptation of Morphology and Control using Graph Neural Networks

Abstract: The co-adaptation of robot morphology and behaviour becomes increasingly important with the advent of fast 3D-manufacturing methods and efficient deep reinforcement learning algorithms. A major challenge for the application of coadaptation methods to the real world is the simulation-to-reality-gap due to model and simulation inaccuracies. However, prior work focuses primarily on the study of evolutionary adaptation of morphologies exploiting analytical models and (differentiable) simulators with large populati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(4 citation statements)
references
References 38 publications
0
4
0
Order By: Relevance
“…Similarly, Ha (2019) proposes to use REINFORCE to optimize policy parameters and design parameters of a population of agents in a joint manner. The co-adaptation method presented by Luck, Amor, and Calandra (2020) improves data-efficiency compared to return-based algorithms by utilizing the critic learned by Soft Actor Critic (SAC) (Haarnoja et al 2018) to query for the expected episodic return of unseen designs during design optimization. While the method we present is closest to the former approach, all discussed co-adaptation methods require access to a reward function, and are thus not capable of co-adapting the behaviour and design of an agent without requiring an engineer to formulate a reward function.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations
“…Similarly, Ha (2019) proposes to use REINFORCE to optimize policy parameters and design parameters of a population of agents in a joint manner. The co-adaptation method presented by Luck, Amor, and Calandra (2020) improves data-efficiency compared to return-based algorithms by utilizing the critic learned by Soft Actor Critic (SAC) (Haarnoja et al 2018) to query for the expected episodic return of unseen designs during design optimization. While the method we present is closest to the former approach, all discussed co-adaptation methods require access to a reward function, and are thus not capable of co-adapting the behaviour and design of an agent without requiring an engineer to formulate a reward function.…”
Section: Related Workmentioning
confidence: 99%
“…Figure 5 shows the results in the two HalfCheetah morphology transfer scenarios. To address RQ3, we compare CoIL to two other co-imitation approaches: using the cheetah without morphology adaptation, as well as to using the Q-function method adapted from Luck, Amor, and Calandra (2020). Since this method is designed for the standard reinforcement learning setting, we adapt it to the imitation learning scenario by using SAIL to imitate the expert trajectories, and iteratively optimizing the morphology using the Q-function.…”
Section: Co-imitation From Simulated Agentsmentioning
confidence: 99%
See 2 more Smart Citations