2022
DOI: 10.1038/s41598-022-05539-7
|View full text |Cite
|
Sign up to set email alerts
|

Federated learning and differential privacy for medical image analysis

Abstract: The artificial intelligence revolution has been spurred forward by the availability of large-scale datasets. In contrast, the paucity of large-scale medical datasets hinders the application of machine learning in healthcare. The lack of publicly available multi-centric and diverse datasets mainly stems from confidentiality and privacy concerns around sharing medical data. To demonstrate a feasible path forward in medical image imaging, we conduct a case study of applying a differentially private federated lear… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
84
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 163 publications
(84 citation statements)
references
References 27 publications
0
84
0
Order By: Relevance
“…Homomorphic encryption 58 , secure multiparty compute 59 , and trusted execution environments (TEEs) 60 , 61 allow for collaborative computations to be performed with untrusted parties while maintaining confidentiality of the inputs to the computation. Differentially private training algorithms 62 – 64 allow for mitigation of information leakage from both the collaborator model updates and the global consensus aggregated models. Finally, assurance that remote computations are executed with integrity can be designed for with the use of hardware-based trust provided by TEEs, as well as with some software-based integrity checking 65 .…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Homomorphic encryption 58 , secure multiparty compute 59 , and trusted execution environments (TEEs) 60 , 61 allow for collaborative computations to be performed with untrusted parties while maintaining confidentiality of the inputs to the computation. Differentially private training algorithms 62 – 64 allow for mitigation of information leakage from both the collaborator model updates and the global consensus aggregated models. Finally, assurance that remote computations are executed with integrity can be designed for with the use of hardware-based trust provided by TEEs, as well as with some software-based integrity checking 65 .…”
Section: Discussionmentioning
confidence: 99%
“…This study is meant to be used as an example for future FL studies between collaborators with an inherent amount of trust that can result in clinically deployable ML models. Further research is required to assess privacy concerns in a detailed manner 63 , 64 and to apply FL to different tasks and data types 66 – 69 . Building on this study, a continuous FL consortium would enable downstream quantitative analyses with implications for both routine practice and clinical trials, and most importantly, increase access to high-quality precision care worldwide.…”
Section: Discussionmentioning
confidence: 99%
“…Perhaps the most known and commonly used FL algorithm is FedAvg which was initially proposed by McMahan et al 13 and then extended for medical image analysis recently by Lu et al 20 and Adnan et al 24 . It learns a global model by aggregating local models trained on independent identically distributed data, as shown in Fig.…”
Section: Methodsmentioning
confidence: 99%
“…Also, we train a centralized model following the traditional way by collecting all training data at the server. For the federated learning paradigm, we use Federated Learning Average (FedAvg) 20 , 24 as the baseline model to compare with. At the evaluation stage, all developed modes are evaluated with the global test set for a fair comparison.…”
Section: Methodsmentioning
confidence: 99%
“…Differential privacy in federated learning is often achieved using differentially-private stochastic gradient descent (DP-SGD) [ 7 , 41 , 42 ], an algorithm that determines the appropriate noise scale and how to clip the model parameter. The combination of federated learning and differential privacy has been explored in multiple medical use cases, including prediction of mortality and adverse drug reactions from electronic health records [ 43 ], brain tumor segmentation [ 9 ], classification of pathology whole slide images [ 20 ], detection of diabetic retinopathy in images of the retina [ 44 ], and identification of lung cancer in histopathologic images [ 45 ].…”
Section: Related Workmentioning
confidence: 99%