2019
DOI: 10.1007/978-3-030-20870-7_6
|View full text |Cite
|
Sign up to set email alerts
|

Robust Deep Multi-modal Learning Based on Gated Information Fusion Network

Abstract: The goal of multi-modal learning is to use complimentary information on the relevant task provided by the multiple modalities to achieve reliable and robust performance. Recently, deep learning has led significant improvement in multi-modal learning by allowing for fusing high level features obtained at intermediate layers of the deep neural network. This paper addresses a problem of designing robust deep multimodal learning architecture in the presence of the modalities degraded in quality. We introduce deep … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
41
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 47 publications
(41 citation statements)
references
References 35 publications
0
41
0
Order By: Relevance
“…First, how to fuse the information. Sophisticated approaches use so-called gating networks which compute averaging weights in inference times, e.g., [ 24 , 25 ]. Another approach is to use recurrent networks to merge autoencoders [ 19 ].…”
Section: Related Workmentioning
confidence: 99%
“…First, how to fuse the information. Sophisticated approaches use so-called gating networks which compute averaging weights in inference times, e.g., [ 24 , 25 ]. Another approach is to use recurrent networks to merge autoencoders [ 19 ].…”
Section: Related Workmentioning
confidence: 99%
“…Fusion robustness is a rarely considered topic in the multi-modal fusion literature, however, it seems to be an important issue in real-world applications, especially in the robotics field. Only several papers [22,29,[31][32][33][34] took into account the non-nominal conditions of the multi-modal fusion and provided some analysis of fusion robustness to data degradation. Such degradation could occur due to sensor noise, its failure, or unexpected weather conditions.…”
Section: Fusion Robustnessmentioning
confidence: 99%
“…To implement an adaptively weighted fusion of multi-modal information in the intermediate feature levels so that the detector is robust to environmental changes, weighting mechanisms that determines the contribution of features from each modality are studied. Kim et al [47] employs the gated information fusion network to adjust the weights of the features from each modality so that degraded features are suppressed. 3D-CVF [5] develops a gated feature fusion network based on spatial-wise attention to generate the fused feature.…”
Section: Object Detection Based On Multi-modal Fusionmentioning
confidence: 99%