2022
DOI: 10.1109/tgrs.2022.3145483
|View full text |Cite
|
Sign up to set email alerts
|

Lightweight Salient Object Detection in Optical Remote Sensing Images via Feature Correlation

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
32
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 65 publications
(32 citation statements)
references
References 63 publications
0
32
0
Order By: Relevance
“…For ORSI-SOD, Li et al [15] proposed the first lightweight method, i.e., CorrNet. They lightened the vanilla VGG-16 [42] for efficient feature extraction, and adopted the coarse-tofine strategy to detect salient objects in ORSIs with dense lightweight refinement blocks.…”
Section: Lightweight Salient Object Detectionmentioning
confidence: 99%
See 2 more Smart Citations
“…For ORSI-SOD, Li et al [15] proposed the first lightweight method, i.e., CorrNet. They lightened the vanilla VGG-16 [42] for efficient feature extraction, and adopted the coarse-tofine strategy to detect salient objects in ORSIs with dense lightweight refinement blocks.…”
Section: Lightweight Salient Object Detectionmentioning
confidence: 99%
“…Saliency maps produced by four types of methods and our method on ORSIs. PA-KRN [12] is an NSI-SOD method, HVPNet [13] is a lightweight NSI-SOD method, MCCNet [14] is an ORSI-SOD method, and CorrNet [15] is a lightweight ORSI-SOD method. Please zoom-in for details.…”
Section: Introductionmentioning
confidence: 99%
See 1 more Smart Citation
“…And cross-modal RGB and TIR features are complementary, which is more conducive to locating objects. Inspired by the object segmentation works which explore the correlation (i.e., co-attention [61]) of target objects in consecutive video frames [36] (i.e., cross-frame) and features of successive levels [37] (i.e., cross-level), we propose Collaborative Location Module to model the pixel-level correlation of cross-modal high-level semantic features to collaboratively determine the object location. In addition, we mine the valuable representation from two types of feature combinations (i.e., summation and multiplication) in the CLM.…”
Section: B Collaborative Location Modulementioning
confidence: 99%
“…Specifically, we divide the basic cross-modal features extracted by the backbone into three levels, i.e., high, middle, and low levels. For high-level features, we propose a Collaborative Location Module (CLM), which builds pixellevel correlations [36], [37] among semantic representations to locate all potential objects. For middle-level features, we propose a Complementary Activation Module (CAM), which builds on the attention mechanism [12], [38] to generate informative features for exact region activation.…”
Section: Introductionmentioning
confidence: 99%