2022
DOI: 10.1109/tgrs.2022.3196771
|View full text |Cite
|
Sign up to set email alerts
|

LESSFormer: Local-Enhanced Spectral-Spatial Transformer for Hyperspectral Image Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
2

Relationship

1
7

Authors

Journals

citations
Cited by 39 publications
(4 citation statements)
references
References 50 publications
0
4
0
Order By: Relevance
“…Sun et al [16] introduce the SSFTT, which incorporates a Gaussian weighted token module into the transformer architecture to extract highlevel semantic features, leading to improved performance. Zou et al [19] design the LESSFormer, including the HSI2Token module and a local-enhanced transformer encoder. This design adaptively generates spectral-spatial tokens and concurrently enhances local information while preserving global context.…”
Section: A Deep Learning-based Methods For Hsi Classificationmentioning
confidence: 99%
“…Sun et al [16] introduce the SSFTT, which incorporates a Gaussian weighted token module into the transformer architecture to extract highlevel semantic features, leading to improved performance. Zou et al [19] design the LESSFormer, including the HSI2Token module and a local-enhanced transformer encoder. This design adaptively generates spectral-spatial tokens and concurrently enhances local information while preserving global context.…”
Section: A Deep Learning-based Methods For Hsi Classificationmentioning
confidence: 99%
“…For the supervised segmentation loss, we employ the conventional cross-entropy loss function [30], [31] to minimize the disparity between the segmentation results and their corresponding reference labels, i.e.,…”
Section: Loss Functionmentioning
confidence: 99%
“…On the one hand, CNNs utilizing local extraction and global parameter sharing mechanisms focus more on spatial content information, potentially losing spectral diagnostic features to some extent [39]. On the other hand, transformer based on global selfattention can effectively model long-range dependencies but often overlook the spatial priors near pixels, which are often more relevant than distant pixels [40]. Therefore, the hybrid structure of CNN and transformer enables the model to capture local spatial information while describing both spatial longrange dependencies and spectral correlations, enhancing the representational capacity of features.…”
Section: Introductionmentioning
confidence: 99%