2021
DOI: 10.1007/978-3-030-69535-4_44
|View full text |Cite
|
Sign up to set email alerts
|

V2A - Vision to Action: Learning Robotic Arm Actions Based on Vision and Language

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 17 publications
0
4
0
Order By: Relevance
“…The performance drop of Mod. [5] and HiTUT [24] is 5.6% and 7.4% lower than the proposed method, SGL. Mod.…”
Section: Outcomes and Analysis For Comparisonmentioning
confidence: 69%
See 3 more Smart Citations
“…The performance drop of Mod. [5] and HiTUT [24] is 5.6% and 7.4% lower than the proposed method, SGL. Mod.…”
Section: Outcomes and Analysis For Comparisonmentioning
confidence: 69%
“…2) Connectionist Method: With the significant evolution of connectionist methods in recent years, deep neural networks show impressive strengths of learning semantic and highdimensional features, which improves robustness to various types of input data. Packing everthing into one network, end-to-end learning models [5], [6] are first proposed to directly map natural language and vision to a sequence of low-level actions. The sequence-to-sequence model suffers from the well-known issue of teacher forcing, which leads to the poor performance under test scenarios.…”
Section: A Human Instrution Followingmentioning
confidence: 99%
See 2 more Smart Citations