2019
DOI: 10.1109/lra.2019.2899918
|View full text |Cite
|
Sign up to set email alerts
|

Learning Navigation Behaviors End-to-End With AutoRL

Abstract: We learn end-to-end point-to-point and pathfollowing navigation behaviors that avoid moving obstacles. These policies receive noisy lidar observations and output robot linear and angular velocities. The policies are trained in small, static environments with AutoRL, an evolutionary automation layer around Reinforcement Learning (RL) that searches for a deep RL reward and neural network architecture with large-scale hyper-parameter optimization. AutoRL first finds a reward that maximizes task completion, and th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
157
0
3

Year Published

2019
2019
2022
2022

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 211 publications
(161 citation statements)
references
References 31 publications
1
157
0
3
Order By: Relevance
“…It has three main steps. First, we learn an obstacle-avoiding point to point policy with AutoRL [4]. Next, since the RL policy avoids obstacles, we can use the policy to generate obstacle-aware reachability training samples by repeatedly rolling out the learned policy.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…It has three main steps. First, we learn an obstacle-avoiding point to point policy with AutoRL [4]. Next, since the RL policy avoids obstacles, we can use the policy to generate obstacle-aware reachability training samples by repeatedly rolling out the learned policy.…”
Section: Methodsmentioning
confidence: 99%
“…Here we show that this is not the case when proxy rewards are used. AutoRL uses proxy rewards (shown in Section III-A) since they significantly improve learning performance, especially for tasks with sparse learning sigals such as navigation [4]. Fig 6a shows examples of two Asteroid trajectories and Fig.…”
Section: E Physical Robot Experimentsmentioning
confidence: 99%
See 3 more Smart Citations