2018
DOI: 10.1177/0278364918760992
|View full text |Cite
|
Sign up to set email alerts
|

Natural language instructions for human–robot collaborative manipulation

Abstract: This paper presents a dataset of natural language instructions for object reference in manipulation scenarios. It comprises 1582 individual written instructions, which were collected via online crowdsourcing. This dataset is particularly useful for researchers who work in natural language processing, human-robot interaction, and robotic manipulation. In addition to serving as a rich corpus of domain-specific language, it provides a benchmark of image-instruction pairs to be used in system evaluations and uncov… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
14
0
1

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 26 publications
(15 citation statements)
references
References 23 publications
0
14
0
1
Order By: Relevance
“…In the field of human-robot interaction (HRI), computer vision [28,29] and speech recognition [30][31][32] are most commonly used as the primary modes of sensing and interaction. However, there has been a recognition of human touch and tactile HRI as an important mode of physical interaction [33].…”
Section: Overview Of Tactile Sensingmentioning
confidence: 99%
“…In the field of human-robot interaction (HRI), computer vision [28,29] and speech recognition [30][31][32] are most commonly used as the primary modes of sensing and interaction. However, there has been a recognition of human touch and tactile HRI as an important mode of physical interaction [33].…”
Section: Overview Of Tactile Sensingmentioning
confidence: 99%
“…Nyga and Beetz [22] curated data sets that provide a sequence How-To instructions for tasks such as preparing recipes. Others such as Jain et al [23], Scalise et al [24] and Mandlekar et al [25] present simulation environments and data sets for tasks such as learning spatial affordances, situated interaction or learning low-level motor skills. The present data sets possess two limitations that make them less usable for the learning task addressed in this work.…”
Section: Related Workmentioning
confidence: 99%
“…The task of visually grounding referring expressions have been also related to the object manipulation [12], [13], for enabling robots to find the target object to manipulate. [12] proposed a dataset that can be used to train a robot or agent that can pick up the target object referred by human language instructions.…”
Section: B Visually Grounding Referring Expressionsmentioning
confidence: 99%
“…The task of visually grounding referring expressions have been also related to the object manipulation [12], [13], for enabling robots to find the target object to manipulate. [12] proposed a dataset that can be used to train a robot or agent that can pick up the target object referred by human language instructions. [13] developed a system which can distinguish the target object referred by human language instruction when multiple objects belonging to the same class are given.…”
Section: B Visually Grounding Referring Expressionsmentioning
confidence: 99%