2021
DOI: 10.1109/access.2021.3093308
|View full text |Cite
|
Sign up to set email alerts
|

Lightweight Channel Attention and Multiscale Feature Fusion Discrimination for Remote Sensing Scene Classification

Abstract: High-resolution remote sensing image scene classification has attracted widespread attention as a basic earth observation task. Remote sensing scene classification aims to assign specific semantic labels to remote sensing scene images to serve specified applications. Convolutional neural networks are widely used for remote sensing image classification due to their powerful feature extraction capabilities. However, the existing methods have not overcome the difficulties of large-scene remote sensing images of l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
6

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(10 citation statements)
references
References 75 publications
0
9
0
Order By: Relevance
“…In addition, compared with EfficientNetB2 with comparable model parameters, our proposed method has also made great improvement in terms of all evaluation metrics. It is worth noting that the proposed method is also compared with a recent study on multi-scale feature fusion and label smoothing, which is proposed for remote sensing classification and is named LmNet ( Wan et al, 2021 ). LmNet takes pre-trained ResNext50 as backbone and combines channel attention, multi-scale feature fusion and label smoothing.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…In addition, compared with EfficientNetB2 with comparable model parameters, our proposed method has also made great improvement in terms of all evaluation metrics. It is worth noting that the proposed method is also compared with a recent study on multi-scale feature fusion and label smoothing, which is proposed for remote sensing classification and is named LmNet ( Wan et al, 2021 ). LmNet takes pre-trained ResNext50 as backbone and combines channel attention, multi-scale feature fusion and label smoothing.…”
Section: Resultsmentioning
confidence: 99%
“…This characteristic encourages the model to learn in the direction with the greatest difference between the correct label and the wrong label, which means only the loss of the correct label position is calculated in the optimization process of the model. However, when the training data is small, and the inter-class similarity and intra-class differences is relatively large, it may cause the network to be overfitting ( Wan et al, 2021 ). To solve the above problems and inspired by previous study ( Szegedy et al, 2017 ; Müller et al, 2019 ), label smoothing is introduced in this study.…”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…Xu R et al [12] combined two different attention mechanisms, control gate attention and feedback attention, in the main and non-main positions of the network, experiment to prove the validity of the model. Wan H et al [13] designed a multi-scale fusion discrimination framework, which is used to add a lightweight attention mechanism to quickly learn important features in the channel and the ability to dilute edge information, and has a good generalization ability. Wu H et al [14] proposed a self-attention network model with joint loss, using the attention model proposed by ResNet-18 integrated to extract image features and reduce the interference of redundant information in complex images.…”
Section: Related Workmentioning
confidence: 99%
“…Classification performance largely depends on features that accurately represent the scene in the image, and thus the extraction of features that describe an image more accurately has become a primary research focus. Recently, convolutional neural networks (CNNs) have been widely utilized in scene classification because they are capable of extracting high-level semantic feature representations for scene classification 13 15 . However, the spatial relationships between features in HRRS images are complex, and there is a large amount of redundant information.…”
Section: Introductionmentioning
confidence: 99%