2022
DOI: 10.48550/arxiv.2201.05119
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?

Abstract: Despite recent progress made by self-supervised methods in representation learning with residual networks, they still underperform supervised learning on the ImageNet classification benchmark, limiting their applicability in performancecritical settings. Building on prior theoretical insights (Mitrovic et al., 2021) we propose RELICv2 which combines an explicit invariance loss with a contrastive objective over a varied set of appropriately constructed data views. RELICv2 achieves 77.1% top-1 classification acc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
36
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
2
2
1

Relationship

0
5

Authors

Journals

citations
Cited by 14 publications
(36 citation statements)
references
References 25 publications
0
36
0
Order By: Relevance
“…Self-supervised learning (SSL), a family of unsupervised machine learning algorithms, may offer a remedy for the latter issue. Crucially, SSL does not need labeled data but instead requires that the internal network representations belonging to related inputs predict one another [15][16][17][18]. In other words, network representations themselves act as targets for similar inputs.…”
Section: Introductionmentioning
confidence: 99%
“…Self-supervised learning (SSL), a family of unsupervised machine learning algorithms, may offer a remedy for the latter issue. Crucially, SSL does not need labeled data but instead requires that the internal network representations belonging to related inputs predict one another [15][16][17][18]. In other words, network representations themselves act as targets for similar inputs.…”
Section: Introductionmentioning
confidence: 99%
“…For a fair comparison, we adopt the same settings as SimCLR to SogCLR unless noted (main difference is the batch size). It is not our focus to leverage multiple techniques for achieving stateof-the-art performance [31]. We also compare with the CLIP framework for bimodal constrastive learning.…”
Section: Methodsmentioning
confidence: 99%
“…The performance of top-1 linear evaluation accuracy was further improved in several works [1,12,25,5,31]. For example, [1] uses clustering assignment (codes) and prediction probabilities over clusters based on learned features to define a cross-entropy loss, where the clustering assignments of each sample are computed online using the prototype vectors and the learned representations.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations