2021
DOI: 10.48550/arxiv.2109.13398
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unrolling SGD: Understanding Factors Influencing Machine Unlearning

Abstract: Machine unlearning is the process through which a deployed machine learning model forgets about one of its training data points. While naively retraining the model from scratch is an option, it is almost always associated with a large computational effort for deep learning models. Thus, several approaches to approximately unlearn have been proposed along with corresponding metrics that formalize what it means for a model to forget about a data point. In this work, we first taxonomize approaches and metrics of … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
7
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(7 citation statements)
references
References 25 publications
0
7
0
Order By: Relevance
“…Scalability to large deletion sets: Many prior unlearning methods assume tiny deletion sets [11,30,53,61,68]. However, we argue practical applications typically require the deletion of larger subsets of data, necessitating that unlearning procedures should scale well to this setting.…”
Section: Problem Formulationmentioning
confidence: 99%
See 1 more Smart Citation
“…Scalability to large deletion sets: Many prior unlearning methods assume tiny deletion sets [11,30,53,61,68]. However, we argue practical applications typically require the deletion of larger subsets of data, necessitating that unlearning procedures should scale well to this setting.…”
Section: Problem Formulationmentioning
confidence: 99%
“…Second, the prohibitive cost of collecting a representative sample of deep networks makes it infeasible to measure the equivalence of model distributions. Most prior work addresses this by measuring similarity of weights [39,61,68] or outputs [27][28][29]51] between a single model sampled from φ u and φ r . However, measuring similarity between two model instances is not representative of the similarity between the distributions they are drawn from.…”
Section: Against Model Indistinguishabilitymentioning
confidence: 99%
“…Since then further extensions to scenarios where efficient analytic solutions could be found were given [24,9], and an extension to unlearn deep neural networks (DNNs) were proposed [2,13,20,12,11,10]. With the growing field of unlearning, there has emerged two categories of machine unlearning algorithms [23]: exact and approximate unlearning, differing by how unlearning is done, and also how the concept of "unlearning" is understood. Exact unlearning for DNNs is based on retraining.…”
Section: Machine Unlearningmentioning
confidence: 99%
“…Existing work have proposed to use techniques such as membership inference [21] to verify the effectiveness of approximate unlearning [12,1] to show that their approximately unlearned models cannot be easily distinguished from models that are not trained on the data points to be unlearned. Alternatively, others compare the similarity of the approximately unlearned models parameters to exactly unlearned models parameters [10,11,25,23].…”
Section: Machine Unlearningmentioning
confidence: 99%
See 1 more Smart Citation