2021
DOI: 10.1109/lra.2021.3067862
|View full text |Cite
|
Sign up to set email alerts
|

Embodying Pre-Trained Word Embeddings Through Robot Actions

Abstract: We propose a promising neural network model with which to acquire a grounded representation of robot actions and the linguistic descriptions thereof. Properly responding to various linguistic expressions, including polysemous words, is an important ability for robots that interact with people via linguistic dialogue. Previous studies have shown that robots can use words that are not included in the action-description paired datasets by using pre-trained word embeddings. However, the word embeddings trained und… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2
2
2

Relationship

1
5

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 25 publications
0
9
0
Order By: Relevance
“…During the conversion between symbolic actions/states and texts in a 3D simulator environment [5], pre-training on large non-paired data has shown an improved performance in zero-shot settings. For description-from/to-action translation, [2] realized appropriate action generation that accepts unseen words not included in the dataset. They successfully retrofitted the pretrained word embeddings into multimodal representations incorporating the action modality.…”
Section: A Utilization Of Pre-trained Model In Translation Tasksmentioning
confidence: 99%
See 4 more Smart Citations
“…During the conversion between symbolic actions/states and texts in a 3D simulator environment [5], pre-training on large non-paired data has shown an improved performance in zero-shot settings. For description-from/to-action translation, [2] realized appropriate action generation that accepts unseen words not included in the dataset. They successfully retrofitted the pretrained word embeddings into multimodal representations incorporating the action modality.…”
Section: A Utilization Of Pre-trained Model In Translation Tasksmentioning
confidence: 99%
“…The present study also adopts this method, following other studies of bidirectional translation between descriptions and actions [1] [2]. A bi-modal autoencoder model has been proposed for the acquisition of multimodal representations of vision and language [11].…”
Section: B Integration Of Multimodal Representationmentioning
confidence: 99%
See 3 more Smart Citations