2020
DOI: 10.1016/j.robot.2019.103312
|View full text |Cite
|
Sign up to set email alerts
|

Multimodal representation models for prediction and control from partial information

Abstract: Similar to humans, robots benefit from interacting with their environment through a number of different sensor modalities, such as vision, touch, sound. However, learning from different sensor modalities is difficult, because the learning model must be able to handle diverse types of signals, and learn a coherent representation even when parts of the sensor inputs are missing. In this paper, a multimodal variational autoencoder is proposed to enable an iCub humanoid robot to learn representations of its sensor… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 25 publications
(15 citation statements)
references
References 44 publications
0
14
0
Order By: Relevance
“…6). 1 Given a single modality of clean image data, Dreamer is generally able to achieve high rewards on the tasks tested [5], [9].…”
Section: Case Study: Table Wipingmentioning
confidence: 99%
See 4 more Smart Citations
“…6). 1 Given a single modality of clean image data, Dreamer is generally able to achieve high rewards on the tasks tested [5], [9].…”
Section: Case Study: Table Wipingmentioning
confidence: 99%
“…Specifically, MuMMI uses PoE fusion [18], which was previously used in a multi-modal variational autoencoder [3] that was later extended to sequential settings [4]. Multi-modal models have also been adopted in robotics applications, where feature vectors from different modalities are concatenated into a single latent representation [1], [2]. Lately, PoE-based fusion has been applied to multi-modal self-supervised training [8], but unlike MuMMI, the method relies on hand-crafted taskdependent losses.…”
Section: Related Workmentioning
confidence: 99%
See 3 more Smart Citations