2022
DOI: 10.1101/2022.03.23.485485
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

PPML-Omics: a Privacy-Preserving federated Machine Learning method protects patients’ privacy in omic data

Abstract: Modern machine learning models towards various tasks with omic data analysis give rise to threats of privacy leakage of patients involved in those datasets. Despite the advances in different privacy technologies, existing methods tend to introduce too much noise, which hampers model accuracy and usefulness. Here, we built a secure and privacy-preserving machine learning (PPML) system by combining federated learning (FL), differential privacy (DP) and shuffling mechanism. We applied this system to analyze data … Show more

Help me understand this report
View published versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3

Citation Types

0
5
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
2
2

Relationship

5
3

Authors

Journals

citations
Cited by 11 publications
(7 citation statements)
references
References 114 publications
0
5
0
Order By: Relevance
“…Meanwhile, those aforementioned regulations will force involved third parties to take actions immediately. According to the requirements of those regulations, not only the previously authorized data by individuals need to be deleted immediately from hosts' storage systems but also 1 Computer Science Program, Computer, Electrical and Mathematical Sciences the associated information should be removed from DL models trained with those data, because DL models could memorize sensitive information of training data and thus expose individual's privacy under risk [7], [8], [9], [10], [11].…”
Section: Introductionmentioning
confidence: 99%
“…Meanwhile, those aforementioned regulations will force involved third parties to take actions immediately. According to the requirements of those regulations, not only the previously authorized data by individuals need to be deleted immediately from hosts' storage systems but also 1 Computer Science Program, Computer, Electrical and Mathematical Sciences the associated information should be removed from DL models trained with those data, because DL models could memorize sensitive information of training data and thus expose individual's privacy under risk [7], [8], [9], [10], [11].…”
Section: Introductionmentioning
confidence: 99%
“…However, their proposed vision-text models were limited to certain tasks. Meanwhile, for ChatCAD, users need to use ChatGPT’s API to upload text descriptions, which could raise data privacy issues [41], [53], [54] as both medical images and text descriptions contain a lot of patients’ private information [55], [56], [57], [58]. To address those issues, MiniGPT-4 [59] is the first open-source method that allows users to deploy locally to interface images with state-of-the-art LLMs and interact using natural language without the need to fine-tune both pre-trained large models but only a small alignment layer.…”
Section: Introductionmentioning
confidence: 99%
“…scPrivacy utilized federated deep metric learning algorithms to train the federated cell type identification model on multiple institutional datasets in data privacy protection manner [ 6 ]. PPML-Omics [ 7 ] analyzed data from three sequencing technologies with a privacy-preserving federated framework, clustering cell populations with Auto-encoder and k-means clustering. Those approaches align with the growing emphasis on ensuring data privacy in bioinformatics, as underscored by the challenges and concerns presented in the field.…”
Section: Introductionmentioning
confidence: 99%