2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS 2021
DOI: 10.1109/igarss47720.2021.9553741
|View full text |Cite
|
Sign up to set email alerts
|

Multi-Modal Self-Supervised Representation Learning for Earth Observation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
8
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3

Relationship

1
6

Authors

Journals

citations
Cited by 12 publications
(8 citation statements)
references
References 7 publications
0
8
0
Order By: Relevance
“…Another work by Chen & Bruzzone, concatenates MS and SAR temporal images to obtain pixel-level discrimination embedding with the distillation network for change detection [24]. Meanwhile, work by Jain et al, demonstrates that single channel features can also provide significant advantages in learning invariant representations for satellite data [3]. Similar results are shown by Montanaro et al [25], which uses a combination of contrastive learning and the text task with the single-channel feature learning.…”
Section: B Ssl and Remote Sensingmentioning
confidence: 77%
See 2 more Smart Citations
“…Another work by Chen & Bruzzone, concatenates MS and SAR temporal images to obtain pixel-level discrimination embedding with the distillation network for change detection [24]. Meanwhile, work by Jain et al, demonstrates that single channel features can also provide significant advantages in learning invariant representations for satellite data [3]. Similar results are shown by Montanaro et al [25], which uses a combination of contrastive learning and the text task with the single-channel feature learning.…”
Section: B Ssl and Remote Sensingmentioning
confidence: 77%
“…Similar results are shown by Montanaro et al [25], which uses a combination of contrastive learning and the text task with the single-channel feature learning. Though it has been seen that similarity-based networks such as SimSiam do not perform well in the case of remote sensing [3], [25], we believe that distillation networks provide extra advantages for the learning of invariant spectral and spatial representations.…”
Section: B Ssl and Remote Sensingmentioning
confidence: 80%
See 1 more Smart Citation
“…An improved version [7] described Ace-Net and included a discriminator at latent space level. Another representation learning framework has been detailed in [26] for a classification task. They learnt a common representation by extracting modality specific features with encoders and projecting one feature space to the other using an additional network.…”
Section: Related Workmentioning
confidence: 99%
“…Contrastive Sensor Fusion [30] constructs positive pairs by using co-located images generated by combining different sensor channels. [31,32,33] apply off-the-shelf contrastive learning algorithms to satellite images. [33] utilizes image inpainting and transformation prediction as additional pretext tasks.…”
Section: Introductionmentioning
confidence: 99%