2018
DOI: 10.1016/j.neunet.2018.07.006
|View full text |Cite
|
Sign up to set email alerts
|

State representation learning for control: An overview

Abstract: Representation learning algorithms are designed to learn abstract features that characterize data. State representation learning (SRL) focuses on a particular kind of representation learning where learned features are in low dimension, evolve through time, and are influenced by actions of an agent. The representation is learned to capture the variation in the environment generated by the agent's actions; this kind of representation is particularly suitable for robotics and control scenarios. In particular, the… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

2
172
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 252 publications
(174 citation statements)
references
References 39 publications
2
172
0
Order By: Relevance
“…In the context of reinforcement learning (which we go into more detail in Sec. III), a good representation encodes the essential information of the state for the agent to choose its next action for a given task [40]. A compact and low-dimensional state representation can make reinforcement learning more data efficient.…”
Section: B Representation Learning For Policy Learningmentioning
confidence: 99%
See 1 more Smart Citation
“…In the context of reinforcement learning (which we go into more detail in Sec. III), a good representation encodes the essential information of the state for the agent to choose its next action for a given task [40]. A compact and low-dimensional state representation can make reinforcement learning more data efficient.…”
Section: B Representation Learning For Policy Learningmentioning
confidence: 99%
“…A popular representation learning objective is reconstruction of the raw sensory input through variational autoencoders [11,29,40,70], which we consider as a baseline in this work. This unsupervised objective benefits learning stability and speed, but it is also data intensive and prone to overfitting [11].…”
Section: B Representation Learning For Policy Learningmentioning
confidence: 99%
“…Following (Lesort et al, 2018), a good state representation should be (1) Markovian (i.e., the current state summarizes all the necessary information to choose an action), (2) able to represent the robot context well enough for policy improvement, (3) able to generalize the learned value-function to unseen states with similar features, and (4), low dimensional for efficient estimation (Böhmer et al, 2015). State representation learning approaches learn low dimensional representations without direct supervision, i.e., exploiting sequences of observations, actions, rewards and generic learning objectives (Lesort et al, 2018).…”
Section: Conceptual Framework and Basic Definitionsmentioning
confidence: 99%
“…State representation learning approaches learn low dimensional representations without direct supervision, i.e., exploiting sequences of observations, actions, rewards and generic learning objectives (Lesort et al, 2018). …”
Section: Conceptual Framework and Basic Definitionsmentioning
confidence: 99%
“…These methods typically learn a nonlinear embedding (e.g., via autoencoders [29,32,26,27]), and-inspired by Koopman operator theory-learn a dynamics model that is constrained to be linear. In a control [24] or reinforcementlearning context [4], the embedding and dynamics models can be learned simultaneously from observations of the state, but most approaches restrict the dynamics to be locally linear [14,19,33,2]. Preprint. The second class of methods corresponds to projection-based dynamics learning (often referred to as "model reduction"), which learns the embedding in a data-driven manner, but computes the dynamics model via a projection process executed on the governing system of ODEs (which must be explicitly known).…”
Section: Introductionmentioning
confidence: 99%