Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP) 2020
DOI: 10.18653/v1/2020.emnlp-main.59
|View full text |Cite
|
Sign up to set email alerts
|

Where Are You? Localization from Embodied Dialog

Abstract: We present WHERE ARE YOU? (WAY), a dataset of ∼6k dialogs in which two humans -an Observer and a Locator -complete a cooperative localization task. The Observer is spawned at random in a 3D environment and can navigate from first-person views while answering questions from the Locator. The Locator must localize the Observer in a detailed top-down map by asking questions and giving instructions. Based on this dataset, we define three challenging tasks: Localization from Embodied Dialog or LED (localizing the Ob… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
15
0
1

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
1

Relationship

1
7

Authors

Journals

citations
Cited by 22 publications
(16 citation statements)
references
References 24 publications
0
15
0
1
Order By: Relevance
“…However, to the best of our knowledge, these prerequisites have not yet been combined with language and the self in mind. Overall, there is a lack of research methods that regard the self in the area of RL, explicitly making use of language in embodied dialogs (Hahn et al, 2020). This section reviews methods that partly satisfy the requirements but still miss at least one of the desired components.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…However, to the best of our knowledge, these prerequisites have not yet been combined with language and the self in mind. Overall, there is a lack of research methods that regard the self in the area of RL, explicitly making use of language in embodied dialogs (Hahn et al, 2020). This section reviews methods that partly satisfy the requirements but still miss at least one of the desired components.…”
Section: Methodsmentioning
confidence: 99%
“…However, the authors do not consider language either. We address this gap by examining the challenges of embodied dialogs (Hahn et al, 2020) in the context of the self, combining the presence of language with other input modalities to learn appropriate hierarchical representations.…”
Section: Scientific Rationale and Contribution Of This Reviewmentioning
confidence: 99%
“…Rather, the oracle describes the target scene and let the navigator to find it, which commonly occurs when someone get lost in a new building. Hahn et al [200] propose a LED task (localizing the Observer from dialog history) to realize such a scenario. Based on this scenario, they present a dataset named WHERE ARE YOU [200] that consists of 6k dialogs of two humans.…”
Section: Vdn: Find a Tablementioning
confidence: 99%
“…Hahn et al [200] propose a LED task (localizing the Observer from dialog history) to realize such a scenario. Based on this scenario, they present a dataset named WHERE ARE YOU [200] that consists of 6k dialogs of two humans. Due to the wide application of multi-agent communication systems [201], [202], [203], [204] in real-world, researchers become interested in implementing dialog navigating in physical environments.…”
Section: Vdn: Find a Tablementioning
confidence: 99%
“…The release of high-quality 3D building and street captures (Chang et al, 2017;Mirowski et al, 2019;Mehta et al, 2020;Xia et al, 2018;Straub et al, 2019) has galvanized interest in developing embodied navigation agents that can operate in complex human environments. Based on these environments, annotations have been collected for a variety of tasks including navigating to a particular class of object (ObjectNav) , navigating from language instructions aka visionand-language navigation (VLN) (Anderson et al, 2018b;Qi et al, 2020;Ku et al, 2020), and vision-and-dialog navigation (Thomason et al, 2020;Hahn et al, 2020). To date, most of these data collection efforts have required the development of custom annotation tools.…”
Section: Introductionmentioning
confidence: 99%