Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction 2020
DOI: 10.1145/3319502.3374824
|View full text |Cite
|
Sign up to set email alerts
|

Teaching a Robot Tasks of Arbitrary Complexity via Human Feedback

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
5

Citation Types

0
5
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1
1

Relationship

0
8

Authors

Journals

citations
Cited by 8 publications
(5 citation statements)
references
References 12 publications
0
5
0
Order By: Relevance
“…Alternative decompositions include subtask sequences called sketches [2], which are less expressive than RMs (the subtasks are run in a single order) and are not hierarchically composable. Context-free grammars defining a subset of English [12], formal languages [26,33,53] and logic-based algebras [39] have been also used to model task composability.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Alternative decompositions include subtask sequences called sketches [2], which are less expressive than RMs (the subtasks are run in a single order) and are not hierarchically composable. Context-free grammars defining a subset of English [12], formal languages [26,33,53] and logic-based algebras [39] have been also used to model task composability.…”
Section: Related Workmentioning
confidence: 99%
“…The so-called 'online' method by Matiisen et al [36] also keeps an estimate of the average return of the different tasks, but it is not presented in an HRL scenario. Wang et al [53] introduce a method that initially learns linear temporal logic formulas for simple tasks, and progressively switches to harder ones leveraging previously learned formulas using a set of predefined templates. Deriving HRMs from (learned) templates is an interesting direction for future work.…”
Section: Related Workmentioning
confidence: 99%
“…Other work [8], [15], [16] has also considered grounding natural language to LTL task specifications with improved accuracy using example trajectories and human-feedback, using formal verification and optimization methods. Patel et al [15] introduced a semi-supervised method for grounding language to LTL via generative models and formal verification, considering whether candidate LTL satisfies trajectories.…”
Section: Related Workmentioning
confidence: 99%
“…Patel et al [15] introduced a semi-supervised method for grounding language to LTL via generative models and formal verification, considering whether candidate LTL satisfies trajectories. Wang et al [16] introduced an algorithm for learning LTL task specifications of arbitrary complexity from human-feedback by a process of eliminating sub-optimal trajectories. Danas et al [8] contributed a three step process for recovering from language grounding errors by performing beam search within a Seq2Seq model to determine the top most likely LTL formulae, differentiating trajectories via maximal semantic differencing, and finally requesting human feedback to clarify which trajectories match the desired task specification.…”
Section: Related Workmentioning
confidence: 99%
“…Another line of work is to consider human prior knowledge of task decomposition to achieve a form of curriculum learning for more complex tasks (Wang et al, 2020 ). Human input to RL has also been used in combination with policy search methods and to improve robot skills on a trajectory level (Celemin and Ruiz-del Solar, 2016 , 2019 ; Celemin et al, 2019 ).…”
Section: Related Workmentioning
confidence: 99%