2019
DOI: 10.48550/arxiv.1911.00357
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DD-PPO: Learning Near-Perfect PointGoal Navigators from 2.5 Billion Frames

Erik Wijmans,
Abhishek Kadian,
Ari Morcos
et al.

Abstract: We present Decentralized Distributed Proximal Policy Optimization (DD-PPO), a method for distributed reinforcement learning in resource-intensive simulated environments. DD-PPO is distributed (uses multiple machines), decentralized (lacks a centralized server), and synchronous (no computation is ever 'stale'), making it conceptually simple and easy to implement. In our experiments on training virtual robots to navigate in Habitat-Sim (Savva et al., 2019), DD-PPO exhibits near-linear scaling -achieving a speedu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
50
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
2
1

Relationship

1
8

Authors

Journals

citations
Cited by 26 publications
(50 citation statements)
references
References 31 publications
(70 reference statements)
0
50
0
Order By: Relevance
“…r t encourages the agent to take actions that help maximize the next separation at each step. We train the policy using Decentralized Distributed PPO (DD-PPO) [75] with trajectory rollouts of 20 steps. The DD-PPO loss consists of a value network loss, policy network loss, and entropy loss to promote exploration (see Supp.…”
Section: Audio-visual Motion Policymentioning
confidence: 99%
“…r t encourages the agent to take actions that help maximize the next separation at each step. We train the policy using Decentralized Distributed PPO (DD-PPO) [75] with trajectory rollouts of 20 steps. The DD-PPO loss consists of a value network loss, policy network loss, and entropy loss to promote exploration (see Supp.…”
Section: Audio-visual Motion Policymentioning
confidence: 99%
“…Robotic navigation. Robots often use visual signals to navigate in novel environments [64,65,66,67]. While vision is often a reliable cue for depth estimation, there are many situations where it is unavailable (e.g.…”
Section: Static Motionmentioning
confidence: 99%
“…Although in the machine learning community there have been several works related to visual navigation in either game-like domains [16], [10] or robotic environments [17], [18], the robotics community has been reluctant to adopt these systems, believing classic methods to perform better for generic tasks [19]. In this context, [5] shows that with sufficient training, RL-based methods can outperform classic methods on fair settings, and [20] shows that RL agents can solve visual navigation tasks with an almost perfect SPL score [15] in simulations. We believe this to be the right direction, but there are still several limitations that need to be addressed, as we show in Section IV.…”
Section: Related Work a Visual Navigationmentioning
confidence: 99%