2020
DOI: 10.3233/jcs-191362
|View full text |Cite
|
Sign up to set email alerts
|

Overfitting, robustness, and malicious algorithms: A study of potential causes of privacy risk in machine learning

Abstract: Machine learning algorithms, when applied to sensitive data, pose a distinct threat to privacy. A growing body of prior work demonstrates that models produced by these algorithms may leak specific private information in the training data to an attacker, either through the models' structure or their observable behavior. This article examines the factors that can allow a training set membership inference attacker or an attribute inference attacker to learn such information. Using both formal and empirical analys… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
41
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
3
1
1

Relationship

1
7

Authors

Journals

citations
Cited by 33 publications
(43 citation statements)
references
References 40 publications
2
41
0
Order By: Relevance
“…While adversarial training produces models that are more invariant to small changes in their inputs, these results show that the training procedure itself can be unstable. This may be related to prior work demonstrating that adversarially-trained models are more vulnerable to membership inference [54,63], a privacy attack that exploits memorization to leak information about training data. While membership vulnerability does not necessarily imply greater LUF, these experiments show that in many cases the two phenomena may be related.…”
Section: Luf and Robust Classificationmentioning
confidence: 69%
See 1 more Smart Citation
“…While adversarial training produces models that are more invariant to small changes in their inputs, these results show that the training procedure itself can be unstable. This may be related to prior work demonstrating that adversarially-trained models are more vulnerable to membership inference [54,63], a privacy attack that exploits memorization to leak information about training data. While membership vulnerability does not necessarily imply greater LUF, these experiments show that in many cases the two phenomena may be related.…”
Section: Luf and Robust Classificationmentioning
confidence: 69%
“…Instability also worsens concrete privacy attacks: oversensitivity to the training set can affect a model's parameters, which can be leveraged to perform membership inference [38,53,61]. Our experiments in Section 5 may suggest that this phenomenon has a connection to leave-one-out unfairness, in that adversarial training increases both LUF and the potential for membership inference attacks [54,63].…”
Section: Related Workmentioning
confidence: 98%
“…In general, three types of prediction error including bias, variance, and irreducible error (noise) are reported in application of individual ML algorithms. [ 53 ]. Therefore, ensemble algorithms were built to improve robustness over a single model with combining the predictions of several models [ 54 , 55 ].…”
Section: Discussionmentioning
confidence: 99%
“…However, despite the advantages of using ML for answering biological questions, possible issues must be overcome in the ML model development and implementation. Overfitting is a common issue when developing ML models, 15 whereby a ML model does not generalize well from observed to unseen data. In this instance, while the model may perform well when making predictions on training data, predictions are not accurate when exposed to new data.…”
Section: Introductionmentioning
confidence: 99%