2022
DOI: 10.48550/arxiv.2204.08499
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DeepCore: A Comprehensive Library for Coreset Selection in Deep Learning

Abstract: Coreset selection, which aims to select a subset of the most informative training samples, is a long-standing learning problem that can benefit many downstream tasks such as data-efficient learning, continual learning, neural architecture search, active learning, etc. However, many existing coreset selection methods are not designed for deep learning, which may have high complexity and poor generalization ability to unseen representations. In addition, the recently proposed methods are evaluated on models, dat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2023
2023
2023
2023

Publication Types

Select...
2

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(7 citation statements)
references
References 24 publications
(45 reference statements)
0
7
0
Order By: Relevance
“…Moreover, the ignored region can be made arbitrarily large for small enough compression levels. Therefore, we expect that the generalization performance will be affected and that the drop in performance will be amplified with smaller compression levels, regardless of the sample size n. This hypothesis is empirically validated (see [1] and Section 5).…”
Section: Asymptotic Behavior Of Sbpamentioning
confidence: 69%
See 3 more Smart Citations
“…Moreover, the ignored region can be made arbitrarily large for small enough compression levels. Therefore, we expect that the generalization performance will be affected and that the drop in performance will be amplified with smaller compression levels, regardless of the sample size n. This hypothesis is empirically validated (see [1] and Section 5).…”
Section: Asymptotic Behavior Of Sbpamentioning
confidence: 69%
“…In this regime, it has been observed (see e.g. [1]) that most SBPA algorithms underperform random pruning (randomly selected subset) 2 . To understand why this occurs, we analyze the asymptotic behavior of SBPA algorithms and identify some of their properties, particularly in the high compression level regime.…”
Section: Connection With Neural Scaling Lawsmentioning
confidence: 99%
See 2 more Smart Citations
“…Coreset-based methods select a certain proportion of data based on certain metrics [5,15]. Lapedriza et al measure the importance of the sample by the benefits obtained from training the model on the sample [24].…”
Section: Dataset Distillationmentioning
confidence: 99%