2020
DOI: 10.1007/978-3-030-58586-0_14
|View full text |Cite
|
Sign up to set email alerts
|

Efficient Attention Mechanism for Visual Dialog that Can Handle All the Interactions Between Multiple Inputs

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
27
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 38 publications
(27 citation statements)
references
References 35 publications
0
27
0
Order By: Relevance
“…A preliminary version of the method won the ALFRED Challenge 2020 1 . The present version further improved the task success rate in unseen and seen environments to 8.37% and 29.16%, respectively, which are significantly higher than the previously published SOTA (0.39% and 3.98%, respectively) [Shridhar et al, 2020].…”
Section: Introductionmentioning
confidence: 48%
See 4 more Smart Citations
“…A preliminary version of the method won the ALFRED Challenge 2020 1 . The present version further improved the task success rate in unseen and seen environments to 8.37% and 29.16%, respectively, which are significantly higher than the previously published SOTA (0.39% and 3.98%, respectively) [Shridhar et al, 2020].…”
Section: Introductionmentioning
confidence: 48%
“…Several variants of VLN tasks have been proposed. A study [Nguyen et al, 2019] allows the agent to communicate with an adviser using natural language to accomplish a given goal. In a study , the agent placed in an environment attempts to find a specified object by communicating with a human by natural language dialog.…”
Section: Embodied Vision-language Tasksmentioning
confidence: 99%
See 3 more Smart Citations