2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01498
|View full text |Cite
|
Sign up to set email alerts
|

HOP: History-and-Order Aware Pretraining for Vision-and-Language Navigation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 41 publications
(18 citation statements)
references
References 24 publications
0
18
0
Order By: Relevance
“…They explore diverse training strategies [84,83], mine extra supervisory signals from synthesized samples [27,71,28] or auxiliary tasks [83,35,53,93,78], and explore intelligent path planning [39,54,81]. For structured and long-range context modeling, recent solutions were developed with environment map [92,13,21,80], transformer architectures [33,61,48,64,11], and multimodal pretraining [56,31,30,12].…”
Section: Related Workmentioning
confidence: 99%
See 4 more Smart Citations
“…They explore diverse training strategies [84,83], mine extra supervisory signals from synthesized samples [27,71,28] or auxiliary tasks [83,35,53,93,78], and explore intelligent path planning [39,54,81]. For structured and long-range context modeling, recent solutions were developed with environment map [92,13,21,80], transformer architectures [33,61,48,64,11], and multimodal pretraining [56,31,30,12].…”
Section: Related Workmentioning
confidence: 99%
“…Later, [30,31,12] conduct pretraining on abundant web image-captions [30] or synthesized trajectory-instruction pairs [31,12] with different VLN-specific proxy tasks. [11,64] introduce historyaware proxy tasks for more VLN-aligned pretraining.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations