2018
DOI: 10.48550/arxiv.1811.04551
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Learning Latent Dynamics for Planning from Pixels

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
220
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
1
1

Relationship

1
8

Authors

Journals

citations
Cited by 100 publications
(221 citation statements)
references
References 0 publications
1
220
0
Order By: Relevance
“…Rich image representations are useful for many downstream tasks. For example, VAEs are often used for latent representation learning in model-based reinforcement learning [8,10,11] by predicting possible future outcomes of the environment. VQ-VAE's latent variables are used as a means for video generation tasks [46].…”
Section: Related Workmentioning
confidence: 99%
“…Rich image representations are useful for many downstream tasks. For example, VAEs are often used for latent representation learning in model-based reinforcement learning [8,10,11] by predicting possible future outcomes of the environment. VQ-VAE's latent variables are used as a means for video generation tasks [46].…”
Section: Related Workmentioning
confidence: 99%
“…Each value is continuous and bounded. We use an action repeat of 2 to produce a better signal to the model (Hafner et al 2019). We keep the original reward, which incentivizes the agent to drive through as many tiles as possible.…”
Section: E1 Environmentsmentioning
confidence: 99%
“…Image-based environments: We implement PlaNet (Hafner et al 2019), which models the environment transition function using a latent dynamics model with deterministic and stochastic transition states; we refer interested readers to the original paper for details. PlaNet does not provide an uncertainty estimate because it only utilizes a single transition model.…”
Section: E2 Uncertainty Estimatorsmentioning
confidence: 99%
“…On the other hand, several recent works incorporate the action (i.e., control command) of the agents in the latent representation [1, 13, 17-19, 29, 41, 42, 46]. While one can use the convolutional recurrent neural network to embed the entire past observations and actions to yield rich temporal information [1], most works first extract the low-dimensional latent dynamics model from the observation with actionconditioned SSM, and integrate the learned latent model into the agent's policy [29] or vision-based planning [19]. However, these approaches use a simple variational autoencoder to extract the latent state and thus cannot represent entity-wise interaction.…”
Section: Latent Dynamics Model From Visual Sequencementioning
confidence: 99%