2022 International Conference on Robotics and Automation (ICRA) 2022
DOI: 10.1109/icra46639.2022.9812230
|View full text |Cite
|
Sign up to set email alerts
|

Learning Latent Actions without Human Demonstrations

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 11 publications
0
3
0
Order By: Relevance
“…Recently, learning-based methods have been employed to generate personalized mappings to compensate for individual differences [12], [13] and task-specific mappings to leverage latent actions [14], [15]. These learning-based mappings follow manifolds derived from observations and are generally dynamic and non-linear.…”
Section: A Control Coordinate Systems In Telemanipulationmentioning
confidence: 99%
“…Recently, learning-based methods have been employed to generate personalized mappings to compensate for individual differences [12], [13] and task-specific mappings to leverage latent actions [14], [15]. These learning-based mappings follow manifolds derived from observations and are generally dynamic and non-linear.…”
Section: A Control Coordinate Systems In Telemanipulationmentioning
confidence: 99%
“…F3: Corrective Feedback -Here, the user has a trajectory showcasing imperfect behavior as a reference and needs to instruct via an improved strategy. This can be done either implicitly, e.g., by pushing a robot into a correct position (Mehta & Losey, 2022), or explicitly via specifying a better action. Implicit corrections require an additional step translating the corrected trajectory into a sequence of agent actions.…”
Section: Feedback Typesmentioning
confidence: 99%
“…Reinforcement learning from human feedback (RLHF) is a powerful tool to train agents when it is difficult to specify a reward function or when human knowledge can improve training efficiency. Recently, using multiple forms of human feedback for reward modeling has come into focus (Jeon et al, 2020;Ghosal et al, 2023;Ibarz et al, 2018;Bıyık et al, 2022a;Mehta & Losey, 2022). Using diverse sources of information opens up several possibilities: (1) feedback from different sources allows for correcting poten-Interactive Learning with Implicit Human Feedback Workshop at ICML 2023, Honolulu, Hawaii, USA, 2023.…”
Section: Introductionmentioning
confidence: 99%