2014
DOI: 10.1186/s40638-014-0008-1
|View full text |Cite
|
Sign up to set email alerts
|

Learning search polices from humans in a partially observable context

Abstract: Decision making and planning for which the state information is only partially available is a problem faced by all forms of intelligent entities they being either virtual, synthetic or biological. The standard approach to mathematically solve such a decisional problem is to formulate it as a partially observable decision process (POMDP) and apply the same optimisation techniques used in the Markov decision process (MDP). However, applying naively the same methodology to solve MDPs as with POMDPs makes the prob… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
2
0

Year Published

2017
2017
2022
2022

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(3 citation statements)
references
References 17 publications
0
2
0
Order By: Relevance
“…Abbu-Dakka [22] solved the peg-in-hole task by human demonstration using exception strategies. Moreover, De Chambier and Suomalainen [23,24] proposed learning compliance for assembly motion demonstration. The key idea of these methods is to learn from human demonstration and exploration for the distribution in their dynamic model, as Dennis mentioned in his research [25].…”
Section: Data Analysis On Mating Connectormentioning
confidence: 99%
“…Abbu-Dakka [22] solved the peg-in-hole task by human demonstration using exception strategies. Moreover, De Chambier and Suomalainen [23,24] proposed learning compliance for assembly motion demonstration. The key idea of these methods is to learn from human demonstration and exploration for the distribution in their dynamic model, as Dennis mentioned in his research [25].…”
Section: Data Analysis On Mating Connectormentioning
confidence: 99%
“…Whereas both are very popular in robotics in general, with different techniques available to reduce the computational complexity, there is a limited number of in-contact tasks they has been used for; popular examples being [168,169], who used POMDP to handle motions requiring both free space and in-contact motions. POMDP has also been used to leverage contact for localization both in classical tasks [170] and when teaching the robot to search for a goal from a human demonstration [171,172,173].…”
Section: Discrete Representationsmentioning
confidence: 99%
“…In [25], autonomous exploration of the whole workspace is combined with tactile-based object discrimination. In [26], the search policy is learned from human behavior to act in the whole workspace, where we instead focus on the transfer of specialized skills to rapidly program and execute a novel task. The approach in [27] combines in-hand object localization using a tactile sensor array with tactile based manipulation.…”
Section: Related Workmentioning
confidence: 99%