2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019
DOI: 10.1109/cvpr.2019.00689
|View full text |Cite
|
Sign up to set email alerts
|

The Regretful Agent: Heuristic-Aided Navigation Through Progress Estimation

Abstract: I know I came from there. Where should I go next? My estimated confidence decreased. Something went wrong. Let's learn this lesson and go back. Instruction: Exit the room. Walk past the display case and into the kitchen. Stop by the table. 20% 13% 25% 42% 60% 75% 90% 1 st step 1 st step 2 nd 5 th 5 th step 4 th 6 th 7 th Figure 1: Vision-and-Language Navigation task and our proposed regretful navigation agent. The agent leverages the selfmonitoring mechanism [14] through time to decide when to roll back to a p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
137
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 165 publications
(137 citation statements)
references
References 19 publications
0
137
0
Order By: Relevance
“…Moreover, we introduce a self-supervised imitation learning method for exploration in order to explicitly address the generalization issue, which is a problem not well-studied in prior work. Concurrent to our work, [44,23,27,28] studies the VLN tasks from various aspects, and [31] introduces a variant of the VLN task to find objects by requesting language assistance when needed. Note that we are the first to propose to explore unseen environments for the VLN task.…”
Section: Related Workmentioning
confidence: 96%
“…Moreover, we introduce a self-supervised imitation learning method for exploration in order to explicitly address the generalization issue, which is a problem not well-studied in prior work. Concurrent to our work, [44,23,27,28] studies the VLN tasks from various aspects, and [31] introduces a variant of the VLN task to find objects by requesting language assistance when needed. Note that we are the first to propose to explore unseen environments for the VLN task.…”
Section: Related Workmentioning
confidence: 96%
“…2 To compensate, beam search is often used to improve success rates. Recent work, e.g., using search strategies (Ke et al, 2019) or progress monitors (Ma et al, 2019b), has focused on mitigating the cost of computing top-k rollouts.…”
Section: Methodsmentioning
confidence: 99%
“…• RCM+SIL(TRAIN) (Wang et al, 2019): an agent trained with cross-modal grounding locally and globally via reinforcement learning. • REGRETFUL (Ma et al, 2019b): an agent with a trained progress monitor heuristic for search that enables backtracking. • FAST (Ke et al, 2019): an agent which combines global and local knowledge to compare partial trajectories of different lengths, enabling efficient backtrack after a mistake.…”
Section: Baseline Systemsmentioning
confidence: 99%
“…The dataset presents a unique challenge as it not only substitutes virtual environments (e.g., MacMahon et al (2006)) with photo-realistic environments but also describes the paths in the environment using human-annotated instructions (as opposed to formulaic instructions provided by mapping applications e.g., ). A number of methods (Anderson et al, 2018b;Fried et al, 2018;Wang et al, 2018a;Ma et al, 2019a;Wang et al, 2018b;Ma et al, 2019b) have been proposed recently to solve the navigation task described in R2R dataset. All these methods build models for agents that learn to navigate in R2R environment and are trained on the entire R2R dataset as well as the augmented dataset introduced by Fried et al 2018which is generated by a speaker model trained on human-annotated instructions.…”
Section: Related Workmentioning
confidence: 99%