2022
DOI: 10.1109/tmm.2021.3069297
|View full text |Cite
|
Sign up to set email alerts
|

Employing Bilinear Fusion and Saliency Prior Information for RGB-D Salient Object Detection

Abstract: Multi-modal feature fusion and saliency reasoning are two core sub-tasks of RGB-D salient object detection. However, most existing models employ linear fusion strategies (e.g., concatenation) for multi-modal feature fusion and use a simple coarse-to-fine structure for saliency reasoning. Despite their simpleness, they can neither fully capture the cross-modal complementary information nor exploit the multi-level complementary information among the cross-modal features at different levels. To address these issu… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 52 publications
(18 citation statements)
references
References 106 publications
0
18
0
Order By: Relevance
“…Furthermore, experimental results on several benchmarks show that, by virtue of the proposed IMFF and LFDF modules, our proposed model can make up the performance drop caused by reducing parameters to some extents. Methods JCUF [39] UCNet [57] DMRA [53] SSF [33] BBSNet [34] BIANet [37] EBFS [32] A2dele [42] DANet [41] CoNet [35] ATSA [40]…”
Section: Discussionmentioning
confidence: 99%
See 2 more Smart Citations
“…Furthermore, experimental results on several benchmarks show that, by virtue of the proposed IMFF and LFDF modules, our proposed model can make up the performance drop caused by reducing parameters to some extents. Methods JCUF [39] UCNet [57] DMRA [53] SSF [33] BBSNet [34] BIANet [37] EBFS [32] A2dele [42] DANet [41] CoNet [35] ATSA [40]…”
Section: Discussionmentioning
confidence: 99%
“…Feature-level fusion based models [20], [22], [32]- [34], [38], [39] first extract the unimodal RGB and depth features from the input RGB and depth images, respectively, and then fuse them to capture their complementary information for SOD. For example, JCUF [39] first employed two subnetwork to extract unimodal RGB and depth features from the input RGB and depth images, respectively, and then designed a multi-branch feature fusion module to jointly use the fused cross-modal features and the unimodal RGB and depth features for SOD.…”
Section: B Rgb-d Sodmentioning
confidence: 99%
See 1 more Smart Citation
“…For RGB-D SOD, our model is compared with several SOTA RGB-D SOD algorithms, including D3Net [78], ASIF-Net [36], ICNet [89], DCMF [52], DRLF [90], SSF [43], SSMA [38], A2dele [46], UC-Net [91], JL-DCF [92], CoNet [44], DANet [81], EBFSP [93],CDNet [94], HAINet [95], RD3D [49], DSA2F [48], MMNet [63] and VST [6]. To ensure the fairness of the comparison results, the saliency maps of the evaluation are provided by the authors or generated by running source codes.…”
Section: Comparisons With Sotas 1) Rgb-d Sodmentioning
confidence: 99%
“…Our model is compared with 16 state-of-the-art RGB-D SOD models, including D3Net [22], ICNet [41], DCMF [6], DRLF [67], SSF [81], SSMA [43], A2dele [57], UCNet [80], CoNet [33], DANet [90], JLDCF [24], EBFSP [31],CDNet [35], HAINet [40], RD3D [10] and DSA2F [61]. To ensure the fairness of the comparison results, the saliency maps of the evaluation are provided by the authors or generated by running source codes.…”
Section: Comparisons With the State-of-the-artmentioning
confidence: 99%