2021
DOI: 10.48550/arxiv.2108.11577
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Machine Unlearning of Features and Labels

Abstract: Removing information from a machine learning model is a non-trivial task that requires to partially revert the training process. This task is unavoidable when sensitive data, such as credit card numbers or passwords, accidentally enter the model and need to be removed afterwards. Recently, different concepts for machine unlearning have been proposed to address this problem. While these approaches are effective in removing individual data points, they do not scale to scenarios where larger groups of features an… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 23 publications
0
4
0
Order By: Relevance
“…Let p(n) be the probability that all P portions are affected on the removal of n samples. Extending the analysis of Warnecke et al [67] from the specific case of SISA to data-influence isolation in general, we get: is a fast increase in the chance of needing a full-retrain as deletion sets get larger. This demonstrates how datainfluence isolation provides little improvement in efficiency compared to the retrain-from-scratch baseline for practical scenarios.…”
Section: D1 Against Isolation Strategiesmentioning
confidence: 88%
See 1 more Smart Citation
“…Let p(n) be the probability that all P portions are affected on the removal of n samples. Extending the analysis of Warnecke et al [67] from the specific case of SISA to data-influence isolation in general, we get: is a fast increase in the chance of needing a full-retrain as deletion sets get larger. This demonstrates how datainfluence isolation provides little improvement in efficiency compared to the retrain-from-scratch baseline for practical scenarios.…”
Section: D1 Against Isolation Strategiesmentioning
confidence: 88%
“…Inexact-unlearning: The primary goal of forgetting can often be hard to achieve, especially in deep networks. Hence, inexact-unlearning literature has relaxed the forgetting goal in two ways: not provable [10,32,35,39,45,56,67] and imperfect [10,31,32,35,39,51,65,68]. Relaxing provability implies unlearning methods do not provide any proven guarantees of information removal.…”
Section: Problem Formulationmentioning
confidence: 99%
“…At lower levels, split variables are selected to greedily maximize splitting conditions such as the Gini index or mutual information. Other methods, including [8,[40][41][42], require parameters to be stored during training, the dwelling of note as their other more salient details we will detail subsequently in other parts.…”
Section: ) In-processing Mumentioning
confidence: 99%
“…Sekari et al [25] studied the difference between differential privacy and machine unlearning. Warnecke et al [26] scaled to the problem of forgetting a group of features and labels from a model. Unlearning for Bayesian methods [27], k-means clustering systems [10], and random forests [8] have also been explored.…”
Section: Related Workmentioning
confidence: 99%