2023
DOI: 10.1109/lgrs.2023.3277347
|View full text |Cite
|
Sign up to set email alerts
|

Spatial-Spectral Decoupling Framework for Hyperspectral Image Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(3 citation statements)
references
References 19 publications
0
3
0
Order By: Relevance
“…4, we can conclude that all parameters have an impact on the proposed method. When we explore the influence of the ration k, we fix the dual-window size as (5,7), and the tradeoff parameters are set to (0.1,1,1). Then, we fix the parameter with the best AUC value in the last round and then explore the optimal value of the next parameter.…”
Section: Parameter Analysismentioning
confidence: 99%
See 1 more Smart Citation
“…4, we can conclude that all parameters have an impact on the proposed method. When we explore the influence of the ration k, we fix the dual-window size as (5,7), and the tradeoff parameters are set to (0.1,1,1). Then, we fix the parameter with the best AUC value in the last round and then explore the optimal value of the next parameter.…”
Section: Parameter Analysismentioning
confidence: 99%
“…1 Compared with visible and multispectral imagery, hyperspectral images (HSIs) contain hundreds of continuous spectral bands that provide spectral-spatial information. On this basis, HSI interpretation techniques have been widely applied in many fields, such as classification, [2][3][4][5] change detection, [6][7][8][9] and object detection. [10][11][12] As an important subbranch of hyperspectral target location, hyperspectral anomaly detection (HAD) aims to distinguish objects of interest by capturing subtle spectral differences of ground objects without any prior information or supervision.…”
Section: Introductionmentioning
confidence: 99%
“…Additionally, the extraction of spectral-spatial features can be realized through multichannel fusion, as exemplified by two-channel CNN (TCCNN) [36] and multi-channel CNN (MCCNN) [37]. Furthermore, inspired by the visual attention mechanism, attention techniques have been integrated into CNNs in various recent methods [38,39]. While existing CNN-based methods have made significant performance strides, the convolution operations in CNN models operate on regular square regions, neglecting pixel-to-pixel relationships.…”
Section: Introductionmentioning
confidence: 99%