2021
DOI: 10.1093/jamia/ocaa341
|View full text |Cite
|
Sign up to set email alerts
|

Federated learning improves site performance in multicenter deep learning without data sharing

Abstract: Objective To demonstrate enabling multi-institutional training without centralizing or sharing the underlying physical data via federated learning (FL). Materials and Methods Deep learning models were trained at each participating institution using local clinical data, and an additional model was trained using FL across all of the institutions. Results We fou… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
83
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
8
1

Relationship

2
7

Authors

Journals

citations
Cited by 144 publications
(83 citation statements)
references
References 24 publications
0
83
0
Order By: Relevance
“…Studies that evaluate the performance of predictive models at multiple sites often involve objective tasks that are easy to replicate, such as image processing or analysis of laboratory tests 25 , 26 , 28 , 39 . In this respect, the current study addresses a non-trivial prediction task: the outcome is subjective and provider-dependent, the variables include non-objective findings such as orders given, and the overall admission rate is influenced by the local patient population.…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Studies that evaluate the performance of predictive models at multiple sites often involve objective tasks that are easy to replicate, such as image processing or analysis of laboratory tests 25 , 26 , 28 , 39 . In this respect, the current study addresses a non-trivial prediction task: the outcome is subjective and provider-dependent, the variables include non-objective findings such as orders given, and the overall admission rate is influenced by the local patient population.…”
Section: Discussionmentioning
confidence: 99%
“…Four main approaches are commonly used today to address the challenges of prediction across multiple sites: (1) Training and testing a model at each site separately in order to build site-specific models 11 , 21 24 ; (2) creating a centralized database of data from all sites in order to build a single uniform model 25 , 26 ; (3) applying a Federated Learning (FL) approach in which the model is trained collaboratively at each site, sharing model parameters but not medical data across sites 27 , 28 ; and (4) when customization is not feasible or when little variation exists between sites, a ready-made model that was trained at one or a few sites, can be used across all sites. This last approach is common practice with clinical risk scores such as the Centor risk score for Group-A Streptococcus pharyngitis or the PECARN algorithm for managing children with traumatic brain injury, to name a few 29 31 .…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, distribution discrepancies in training data from these populations result in biases that are one of the major hindrances before generalizing ML approaches. Given the large volume and diverse data needed for model training, Federated Learning (FL) approaches may provide a novel opportunity for the future of ML applications (Rajendran et al, 2021;Sarma et al, 2021). FL is a collaborative ML training approach in which training data is not centralized and stays within organizational boundaries (Figure 1).…”
Section: Introductionmentioning
confidence: 99%
“…Examples include the 100 patient Prostate MR Image Segmentation (PROMISE12) challenge dataset [10] and the 60 patient NCI-ISBI (National Cancer Institute-International Symposium on Biomedical Imaging) Automated Segmentation of Prostate Structures (ASPS13) challenge dataset [11]. Other algorithms have been trained on institutionally developed local datasets that include between 100 and 650 studies [12][13][14][15]. Unfortunately, the development of research-quality prostate boundary annotations is challenging.…”
Section: Introductionmentioning
confidence: 99%