2022
DOI: 10.48550/arxiv.2207.08051
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

SatMAE: Pre-training Transformers for Temporal and Multi-Spectral Satellite Imagery

Abstract: Unsupervised pre-training methods for large vision models have shown to enhance performance on downstream supervised tasks. Developing similar techniques for satellite imagery presents significant opportunities as unlabelled data is plentiful and the inherent temporal and multi-spectral structure provides avenues to further improve existing pre-training strategies. In this paper, we present SatMAE, a pre-training framework for temporal or multi-spectral satellite imagery based on Masked Autoencoder (MAE). To l… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(11 citation statements)
references
References 30 publications
0
10
0
Order By: Relevance
“…WHU Aerial Vaihingen ResNet50 (ImageNet-1k) [21] 88.5 74.0 SeCo [31] 86.7 68.9 ViT (ImageNet-22k) [15] 81.6 72.6 SatMAE [10] 82.5 70.6 Swin (random) [28] 88.2 67.0 Swin (ImageNet-22k) [28] 90.4 74.7 GFM 90.7 75.3…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…WHU Aerial Vaihingen ResNet50 (ImageNet-1k) [21] 88.5 74.0 SeCo [31] 86.7 68.9 ViT (ImageNet-22k) [15] 81.6 72.6 SatMAE [10] 82.5 70.6 Swin (random) [28] 88.2 67.0 Swin (ImageNet-22k) [28] 90.4 74.7 GFM 90.7 75.3…”
Section: Methodsmentioning
confidence: 99%
“…[40] employs a colorization objective on Sentinel-2 imagery utilizing the various spectral bands. Most recently, SatMAE [10] explores the use of masked image modeling to train a large ViT model. This work is similar in some respect to ours, as we also train a transformer model with an MIM objective.…”
Section: Geospatial Pretrainingmentioning
confidence: 99%
See 3 more Smart Citations