2020
DOI: 10.1109/access.2020.2971509
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Modal Weights Sharing and Hierarchical Feature Fusion for RGBD Salient Object Detection

Abstract: Salient object detection (SOD) aims to identify and locate the most attractive regions in an image, which has been widely used in various vision tasks. Recent years, with the development of RGBD sensor technology, depth information of scenes becomes available for image understanding. In this paper, we systematically investigate and evaluate on how to integrate depth cues in a pre-trained deep network and learn informative features for SOD. First, we propose a CNN-based cross-modal transfer learning, which lear… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(2 citation statements)
references
References 41 publications
0
2
0
Order By: Relevance
“…The proposed DCNN aims to overcome the limitations of traditional fusion methods and single CNNbased approaches by effectively capturing the complementary information present in multiple modalities and improving the fusion quality. The ensemble learning framework is expected to leverage the diversity and strengths of individual networks to enhance the accuracy, robustness, and interpretability of the fusion process [22].…”
Section: Problem Statementmentioning
confidence: 99%
“…The proposed DCNN aims to overcome the limitations of traditional fusion methods and single CNNbased approaches by effectively capturing the complementary information present in multiple modalities and improving the fusion quality. The ensemble learning framework is expected to leverage the diversity and strengths of individual networks to enhance the accuracy, robustness, and interpretability of the fusion process [22].…”
Section: Problem Statementmentioning
confidence: 99%
“…A number of RGB-D saliency detection methods focus on the fusion of cross-modal information. Xiao et al [54] employed a CNN-based cross-modal transfer learning framework to guide the depth domain feature extraction. Wang et al [55] designed two-streamed convolutional neural networks to extract features and employed a switch map to adaptively fuse the predicted saliency maps.…”
Section: Rgb-d Saliency Detectionmentioning
confidence: 99%