2021
DOI: 10.48550/arxiv.2106.07760
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RETRIEVE: Coreset Selection for Efficient and Robust Semi-Supervised Learning

Abstract: Semi-supervised learning (SSL) algorithms have had great success in recent years in limited labeled data regimes. However, the current state-of-the-art SSL algorithms are computationally expensive and entail significant compute time and energy requirements. This can prove to be a huge limitation for many smaller companies and academic groups. Our main insight is that training on a subset of unlabeled data instead of entire unlabeled data enables the current SSL algorithms to converge faster, thereby reducing t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 12 publications
0
4
0
Order By: Relevance
“…For instance, one could argue that preserving the gradient is an important feature to have in the coreset as it would lead to similar minima [31,32]. Other work considered the problem of corset selection as a two-stage optimization problem where the subset selection can be seen also as an optimization problem [33,34]. Other methods consider conisder the likelihood and its connection with submodular functions in order to select the subset [35,36].…”
Section: Non-sbpa Methodsmentioning
confidence: 99%
“…For instance, one could argue that preserving the gradient is an important feature to have in the coreset as it would lead to similar minima [31,32]. Other work considered the problem of corset selection as a two-stage optimization problem where the subset selection can be seen also as an optimization problem [33,34]. Other methods consider conisder the likelihood and its connection with submodular functions in order to select the subset [35,36].…”
Section: Non-sbpa Methodsmentioning
confidence: 99%
“…Existing studies usually consider the selection of subset (optimization of samples S or selection weights w) as the outer objective and the optimization of model parameters θ on S as the inner objective. Representative methods include cardinality-constrained bilevel optimization [24] for continual learning, RETRIEVE for semisupervised learning (SSL) [25], and GLISTER [2] for supervised learning and active learning.…”
Section: Bilevel Optimization Based Methodsmentioning
confidence: 99%
“…RETRIEVE. The RETRIEVE method [25] discusses the scenario of SSL under bilevel optimization, where we have both a labeled set T and an unlabled set P. The bilevel optimization problem in RETRIEVE is formulated as…”
Section: Bilevel Optimization Based Methodsmentioning
confidence: 99%
See 1 more Smart Citation