2020 IEEE International Conference on Robotics and Automation (ICRA) 2020
DOI: 10.1109/icra40945.2020.9197393
|View full text |Cite
|
Sign up to set email alerts
|

SA-Net: Robust State-Action Recognition for Learning from Observations

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
17
0

Year Published

2020
2020
2022
2022

Publication Types

Select...
5
4
1

Relationship

1
9

Authors

Journals

citations
Cited by 26 publications
(17 citation statements)
references
References 17 publications
0
17
0
Order By: Relevance
“…Deep learning also opens the possibility of tackling the learning problem in an end-to-end manner. A recent step in this direction is SA-Net (Soans et al 2020), which is able to recognize state-action pairs using an RGB-D sensor. However, it still needs to be coupled with a policy learning algorithm.…”
Section: Mapping Sensors To Actuatorsmentioning
confidence: 99%
“…Deep learning also opens the possibility of tackling the learning problem in an end-to-end manner. A recent step in this direction is SA-Net (Soans et al 2020), which is able to recognize state-action pairs using an RGB-D sensor. However, it still needs to be coupled with a policy learning algorithm.…”
Section: Mapping Sensors To Actuatorsmentioning
confidence: 99%
“…To overcome the obstacle of noisy labels or features, perturbation schemes and data augmentations have been investigated. In computer vision, data augmentation is done by applying operations like cropping and rotation to combat potential mislabelled training data [2,17,12]. Another line of work achieves robustness against noisy data by generating data synthesizers that achieves the same predictive performance as using the real data.…”
Section: Data Augmentation and Synthetic Data Generation For Ro-mentioning
confidence: 99%
“…Various methods exist to tackle the RL optimization problem. Among them, Policy Gradient techniques have enabled to use RL in real-world robotic contexts (see [31] for an extensive definition of these approaches and [32,33,34] for successful real-world robotic applications). These techniques are based on the use of a stochastic policy formally denoted by π θ , where θ is a vector of parameters.…”
Section: Model-free Adaptive Controlmentioning
confidence: 99%