2022
DOI: 10.1117/1.jrs.17.022205
|View full text |Cite
|
Sign up to set email alerts
|

AMFFNet: attention-guided multi-level feature fusion network for land cover classification of remote sensing images

Abstract: In the field of remote sensing, the classification of land cover is a pivotal and challenging issue. Standard models fail to capture global and semantic information in remote sensing images despite the fact that a convolutional neural network provides robust support for semantic segmentation. In addition, owing to disparities in semantic levels and spatial resolution, the simple fusion of low-level and high-level features may diminish the efficiency. To address these deficiencies, an attention-guided multi-lev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(3 citation statements)
references
References 50 publications
0
3
0
Order By: Relevance
“…In order to merge global and local features extracted from both sides smoothly, we propose a decoder similar to a feature pyramid structure that fuses features progressively at different levels from bottom to top. It also enhances the representation of local features at the first layer by connecting it to the contextual path [48] using a double attention module consisting of parallel spatial attention and channel attention. Specifically, the SFF module has five layers, with the first layer using channel attention, spatial attention and two 3x3 standard convolution blocks, the structure of which can be seen in Figure 4, and the remaining layers consisting of two 3x3 standard convolution blocks and one 2x upsampling block.…”
Section: Local Embedding Module and Coordinate Attention Fusion Modulementioning
confidence: 99%
See 2 more Smart Citations
“…In order to merge global and local features extracted from both sides smoothly, we propose a decoder similar to a feature pyramid structure that fuses features progressively at different levels from bottom to top. It also enhances the representation of local features at the first layer by connecting it to the contextual path [48] using a double attention module consisting of parallel spatial attention and channel attention. Specifically, the SFF module has five layers, with the first layer using channel attention, spatial attention and two 3x3 standard convolution blocks, the structure of which can be seen in Figure 4, and the remaining layers consisting of two 3x3 standard convolution blocks and one 2x upsampling block.…”
Section: Local Embedding Module and Coordinate Attention Fusion Modulementioning
confidence: 99%
“…To demonstrate the advanced nature of our method, we conducted a series of comparison experiments on landcover.ai between SABNet and nine other state-of-the-art land cover classification methods. To demonstrate the fairness of the experiments, in addition to the benchmark network BiseNet, we selected network structures with similar parametric numbers as SABNet, namely: UNet [16], Deeplabv3+ [26], PSPNet [17], SegNet [25], DIResUNet [51], FCN-8s [15], AMFFNet [48], HFENet [29], DEANet [52] and Segformer [41], and analysed each network. UNet, DeepLabv3+, SegNet, DIRe-sUNet and AMFFNet represent encoder-decoder networks.…”
Section: F Comparison With Other Advanced Network On the Land-coverai...mentioning
confidence: 99%
See 1 more Smart Citation