2023
DOI: 10.48550/arxiv.2302.04476
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

GFM: Building Geospatial Foundation Models via Continual Pretraining

Abstract: Geospatial technologies are becoming increasingly essential in our world for a large range of tasks, such as earth monitoring and natural disaster response. To help improve the applicability and performance of deep learning models on these geospatial tasks, various works have pursued the idea of a geospatial foundation model, i.e., training networks from scratch on a large corpus of remote sensing imagery. However, this approach often requires a significant amount of data and training time to achieve suitable … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
12
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(12 citation statements)
references
References 40 publications
0
12
0
Order By: Relevance
“…The first line of research aims to learn the representations of remote sensing images through establishing the similarity metrics between multiple images captured at different geolocations or multiple views of the same object/location (Jean et al 2019;Manas et al 2021;Ayush et al 2021). Another line of research attempts to train a vision model with the goal to reconstruct masked patches of the input image (Cong et al 2022;Fuller, Millard, and Green 2022a,b;Reed et al 2022;Mendieta et al 2023;Cha, Seo, and Lee 2023)…”
Section: Related Workmentioning
confidence: 99%
“…The first line of research aims to learn the representations of remote sensing images through establishing the similarity metrics between multiple images captured at different geolocations or multiple views of the same object/location (Jean et al 2019;Manas et al 2021;Ayush et al 2021). Another line of research attempts to train a vision model with the goal to reconstruct masked patches of the input image (Cong et al 2022;Fuller, Millard, and Green 2022a,b;Reed et al 2022;Mendieta et al 2023;Cha, Seo, and Lee 2023)…”
Section: Related Workmentioning
confidence: 99%
“…The three elements that make up a foundation model include a large-scale dataset, a pretrained model, and the pretraining method. Since 2021, several million-scale publicly available large-scale datasets have been built in remote sensing, including FMoW [249], SeCo [250], Million-AID [251], Levir-KR [252], GeoPile [253], SSL4EO-S12 [254] and SSL4EO-L [255]. In the early days, pretrained model research typically deals with ResNet50 [250], [252], ViT Large [256], [257], Swin Base [253], [258], ViTAE-B [259], and ViTAEv2-S [260], the number of parameters of them was relatively small compared those in computer vision.…”
Section: E Pretrained Models For Ssrsimentioning
confidence: 99%
“…Since 2021, several million-scale publicly available large-scale datasets have been built in remote sensing, including FMoW [249], SeCo [250], Million-AID [251], Levir-KR [252], GeoPile [253], SSL4EO-S12 [254] and SSL4EO-L [255]. In the early days, pretrained model research typically deals with ResNet50 [250], [252], ViT Large [256], [257], Swin Base [253], [258], ViTAE-B [259], and ViTAEv2-S [260], the number of parameters of them was relatively small compared those in computer vision. Recently, with the rapid development of vision foundation models, the size of remote sensing pretrained models has expanded to billon-scale [261].…”
Section: E Pretrained Models For Ssrsimentioning
confidence: 99%
See 1 more Smart Citation
“…More broadly, continual pretraining has been found to improve the performance of self-supervised models on many tasks [6]. As RS data is available as large but unlabeled datasets, this has become a popular technique to specialize large-scale models for remote EO [55]. [56], [57] e.g.…”
Section: Continual Pretraining and Knowledge Distillationmentioning
confidence: 99%