2021 IEEE/CVF International Conference on Computer Vision (ICCV) 2021
DOI: 10.1109/iccv48922.2021.01016
|View full text |Cite
|
Sign up to set email alerts
|

Mean Shift for Self-Supervised Learning

Abstract: County (UMBC)ScholarWorks@UMBC digital repository on the Maryland Shared Open Access (MD-SOAR) platform.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
24
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 64 publications
(24 citation statements)
references
References 33 publications
0
24
0
Order By: Relevance
“…Besides, since the combined patch embeddings only contain part of the information in the whole image, pulling the partially combined patches closer to the target view that contains the whole image information is more challenging than pulling the original image pairs and implicitly increasing the asymmetric of the network structure, which have been demonstrated beneficial for increasing the richness of feature representations and improve the self-supervised learning performance [15,11,22]. Owing to these merits, Fast-MoCo can achieve high sample utilization efficiency with marginal extra computational cost and thus obtain promising performance with much less training time.…”
Section: Discussionmentioning
confidence: 99%
“…Besides, since the combined patch embeddings only contain part of the information in the whole image, pulling the partially combined patches closer to the target view that contains the whole image information is more challenging than pulling the original image pairs and implicitly increasing the asymmetric of the network structure, which have been demonstrated beneficial for increasing the richness of feature representations and improve the self-supervised learning performance [15,11,22]. Owing to these merits, Fast-MoCo can achieve high sample utilization efficiency with marginal extra computational cost and thus obtain promising performance with much less training time.…”
Section: Discussionmentioning
confidence: 99%
“…NNCLR [23] compares with SimCLR to select the most similar sample representation from a queue by the nearest neighbor method instead of the original representation to calculate the contrastive loss, thus improving the model performance by increasing the training complexity. MSF [24] compared to SimCLR calculates the contrastive loss by selecting the K most similar sample representations in a queue and calculating the mean instead of the original sample representation. TTL [25] and HardCL [26] sensor data encoder encoder However, with the downstream task identified as a classification task, the above work is contradictory for how negative examples are defined because this definition would consider each instance as a single class.…”
Section: Related Work a Definition Of Contrastive Learning Negative E...mentioning
confidence: 99%
“…The pre-trained learned model is fine-tuned in downstream tasks using a small amount of labeled data to achieve performance comparable to supervised learning [16] [17]. There are many types of pre-training tasks for contrastive learning, such as MoCo [18] [19] and SimCLR [20] [21] with instance discrimination [22] as the task, and NNCLR [23], MSF [24], TTL [25] and HardCL [26] which redefine positive and negative pairs based on the instance discrimination task. In addition to this, SwAV [27] uses clustering to reduce feature dimensionality, BYOL [28] and SimSiam [29] that drop the use of negative examples and use similarity metrics for the pretraining task.…”
Section: Introductionmentioning
confidence: 99%
“…For example, pictures of different cats should not be considered completely negative to each other. To this end, heuristically modifying the inter-sample relations have been widely applied in selfsupervised learning [10,12,22,34], intuitively similar to us. Many of the state-of-the-art methods can be considered as special cases of our method by the adjustment of hyperparameters (Sec.…”
Section: Related Workmentioning
confidence: 99%