Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing 2018
DOI: 10.18653/v1/d18-1286
|View full text |Cite
|
Sign up to set email alerts
|

Translating Navigation Instructions in Natural Language to a High-Level Plan for Behavioral Robot Navigation

Abstract: We propose an end-to-end deep learning model for translating free-form natural language instructions to a high-level plan for behavioral robot navigation. The proposed model uses attention mechanisms to connect information from user instructions with a topological representation of the environment. To evaluate this model, we collected a new dataset for the translation problem containing 11,051 pairs of user instructions and navigation plans. Our results show that the proposed model outperforms baseline approac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 23 publications
(18 citation statements)
references
References 26 publications
0
18
0
Order By: Relevance
“…Shah et al [15] utilized attention over linguistic instructions conditioned on the multi-modal sensory observations to focus on the relevant parts of the command during navigation task. [16] approach the language-based navigation task as a sequence prediction problem. They translate navigation instructions into a sequence of behaviours that a robot can execute to reach the desired destination.…”
Section: Language Based Navigationmentioning
confidence: 99%
“…Shah et al [15] utilized attention over linguistic instructions conditioned on the multi-modal sensory observations to focus on the relevant parts of the command during navigation task. [16] approach the language-based navigation task as a sequence prediction problem. They translate navigation instructions into a sequence of behaviours that a robot can execute to reach the desired destination.…”
Section: Language Based Navigationmentioning
confidence: 99%
“…In the literature, a few datasets have been created for similar tasks Chen et al 2019;de Vries et al 2018;Zang et al 2018). For instance, annotated the language description for the route by asking the user to navigate the entire path in egocentric perspective.…”
Section: Annotation and Dataset Statisticsmentioning
confidence: 99%
“…For instance, annotated the language description for the route by asking the user to navigate the entire path in egocentric perspective. Incorporation of overhead map of navigated route as an aid for describing the route can be seen in Chen et al (2019);de Vries et al (2018); Zang et al (2018).…”
Section: Annotation and Dataset Statisticsmentioning
confidence: 99%
“…Most strategies are based on imitation learning, relying on expert demonstrations and knowledge from the environment. For example, [42] relate instructions to an environment graph, requiring both demonstrations and high-level navigation information. Closer to our work, [15] also learns a navigation model and an instruction generator, but the latter is used to generate additional training data for the agent.…”
Section: A Vision and Language Navigationmentioning
confidence: 99%
“…Besides, instruction following is a notoriously hard RL problem as the training signal is very sparse since the agent is only rewarded over task completion. In practice, the navigation and language grounding problems are often circumvented by warm-starting the policy with labeled trajectories [42,1]. Although scalable, these approaches require numerous human demonstrations, whereas we here want to jointly learn the navigation policy and language understanding from scratch.…”
Section: Introductionmentioning
confidence: 99%