2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01556
|View full text |Cite
|
Sign up to set email alerts
|

Crafting Better Contrastive Views for Siamese Representation Learning

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
18
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
3

Relationship

0
8

Authors

Journals

citations
Cited by 72 publications
(18 citation statements)
references
References 24 publications
0
18
0
Order By: Relevance
“…For example, D w was designed to be large enough for self-supervised learning, which has received surprisingly little attention in the WSOL community [9]. We are also interested in using iNatLoc500 to study whether self-supervised learning methods can be improved by using WSOL methods to select crops [40], especially in the context of fine-grained data [15]. For the object detection community, the clean boxes in iNatLoc500 can (i) serve as a test set for object detectors trained on the noisy iNat17 boxes, (ii) be used to study the problem of learning multi-instance detectors from one box per image, and (iii) be used to analyze the role of label granularity in object detection.…”
Section: Discussionmentioning
confidence: 99%
“…For example, D w was designed to be large enough for self-supervised learning, which has received surprisingly little attention in the WSOL community [9]. We are also interested in using iNatLoc500 to study whether self-supervised learning methods can be improved by using WSOL methods to select crops [40], especially in the context of fine-grained data [15]. For the object detection community, the clean boxes in iNatLoc500 can (i) serve as a test set for object detectors trained on the noisy iNat17 boxes, (ii) be used to study the problem of learning multi-instance detectors from one box per image, and (iii) be used to analyze the role of label granularity in object detection.…”
Section: Discussionmentioning
confidence: 99%
“…This was followed by papers that simplified this approach [13][14][15]37] and proposed non-contrastive variants [16,33]. While those approaches have been suc-cessful, the utility of augmentation-based self-supervised learning has been questioned [65,102] with follow-up work proposing the use of objectness [63,72] and saliency [77] to alleviate some of those concerns. While we share the motivation for visual representation learning, we also question the reliance on image augmentation and propose using language sampling to learn better invariances.…”
Section: Related Workmentioning
confidence: 99%
“…We follow the code implementation and hyper-parameter of [68]. For small datasets (i.e., CIFAR-10/100 and Tiny ImageNet), we use the same training setup in all experiments.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…Since the data size is larger, we train SimCLR and our method with the batch size of 1024 and cosine-annealed learning rate of 0.6 for faster convergence. We adopt the same setting as in [68] for training the linear classifier.…”
Section: Implementation Detailsmentioning
confidence: 99%