2018
DOI: 10.48550/arxiv.1807.06233
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Robust Deep Multi-modal Learning Based on Gated Information Fusion Network

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2019
2019
2019
2019

Publication Types

Select...
3

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 26 publications
0
3
0
Order By: Relevance
“…for combining modalities [27,9,30], and recently it has been shown that variations of attention and gating mechanisms [16,2,32] can adapt to dynamic environments by weighting modalities differently for conditions that occur in the training data. While most works [21,31] have considered external environmental degradations such as rain, snow, glare, low-lighting, and seasonal appearance changes, more recent works [17] address robustness to different types of internal image degradations such as Gaussian noise.…”
Section: Related Work Fusion Architectures For Rgb-d Semantic Segment...mentioning
confidence: 99%
“…for combining modalities [27,9,30], and recently it has been shown that variations of attention and gating mechanisms [16,2,32] can adapt to dynamic environments by weighting modalities differently for conditions that occur in the training data. While most works [21,31] have considered external environmental degradations such as rain, snow, glare, low-lighting, and seasonal appearance changes, more recent works [17] address robustness to different types of internal image degradations such as Gaussian noise.…”
Section: Related Work Fusion Architectures For Rgb-d Semantic Segment...mentioning
confidence: 99%
“…Part to be attention in each modal is acquired by a network itself. For example, in the case of images, the pixel of interest in the image is used [25].…”
Section: Multi-modal Learningmentioning
confidence: 99%
“…2) These multiple intermediate features are merged into a joint representation via a fusion strategy. Such a fusion approach is referred to as intermidiate fusion [11] because the powerful intermediate features obtained by deep neural networks (DNNs) are merged to construct the joint representation. Deep multimodal learning has been shown to achieve remarkable performance for many machine learning tasks, such as deep cross-modal hashing [12] and deep semantic multimodal hashing [13].…”
Section: Introductionmentioning
confidence: 99%