2019
DOI: 10.1162/artl_a_00301
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning for Improving Agent Design

Abstract: In many reinforcement learning tasks, the goal is to learn a policy to manipulate an agent, whose design is fixed, to maximize some notion of cumulative reward. The design of the agent's physical structure is rarely optimized for the task at hand. In this work, we explore the possibility of learning a version of the agent's design that is better suited for its task, jointly with the policy. We propose an alteration to the popular OpenAI Gym framework, where we parameterize parts of an environment, and allow an… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
62
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 82 publications
(63 citation statements)
references
References 66 publications
1
62
0
Order By: Relevance
“…However, their method results in only simple robots with a few components, and assumes physically implausible reconfigurability. Reinforcement learning has also been applied in order to continually improve agent design [Ha 2018]. In their design framework, both the environment and robot are altered to enable the agent to learn more effectively.…”
Section: Generative Design For Roboticsmentioning
confidence: 99%
“…However, their method results in only simple robots with a few components, and assumes physically implausible reconfigurability. Reinforcement learning has also been applied in order to continually improve agent design [Ha 2018]. In their design framework, both the environment and robot are altered to enable the agent to learn more effectively.…”
Section: Generative Design For Roboticsmentioning
confidence: 99%
“…For example, in voxel-based robot simulators, [61,62] which use voxels as structural building blocks, there are 4.5 × 10 8 unique ways to arrange 12 voxels to form a robot, and the design space (the number of possible designs) increases exponentially with each additional block. [63] As a result, evolutionary [64][65][66][67] and learning [68,69] algorithms are usually employed to efficiently explore the vast space of possible robot designs. [70] Directly incorporating biologically inspired mechanisms of shape change-for example, slowly extruding limbs during optimization rather than optimizing controllers only for the final, legged form of the robot-has been shown to speed the evolution of robust, adaptive behavior in simulated robots.…”
Section: Simulated Shape Changing Robotsmentioning
confidence: 99%
“…Deep RL has recently been used for robot design (Schaff et al 2018;Ha 2018) to simultaneously learn a design and a control policy. Deep RL requires only a sparse reward function be formulated for each task.…”
Section: Related Workmentioning
confidence: 99%
“…We compare our approach to two related methods which search for modular arrangements: a best-first search (Ha et al 2018) and an evolutionary search (Icer et al 2017). After training the DQN, our algorithm finds lower-cost solutions more efficiently than these related methods.…”
Section: Introductionmentioning
confidence: 99%