2022
DOI: 10.1109/access.2022.3141709
|View full text |Cite
|
Sign up to set email alerts
|

Extremely Randomized Trees With Privacy Preservation for Distributed Structured Health Data

Abstract: Machine learning has recently attracted a lot of attention in the healthcare domain. The data used by machine learning algorithms in healthcare applications is often distributed over multiple sources, e.g., hospitals. One main difficulty lies in analyzing such data without compromising personal information, which is a primary concern in healthcare applications. Therefore, in these applications, we are interested in running machine learning algorithms over distributed data without disclosing sensitive informati… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
12
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 28 publications
(12 citation statements)
references
References 139 publications
0
12
0
Order By: Relevance
“…The extra-tree classifier uses a meta estimator that fits a number of randomized decision trees (i.e., extra trees) on various subsamples of the dataset and employs averaging to improve the predictive accuracy and control overfitting [71] . This classifier has several parameters, e.g., criterion, maximum of features ( max_features ), minimum of the sample leaf ( min_sample_leaf ), and minimum of the sample split ( min_sample_split ).…”
Section: Resultsmentioning
confidence: 99%
“…The extra-tree classifier uses a meta estimator that fits a number of randomized decision trees (i.e., extra trees) on various subsamples of the dataset and employs averaging to improve the predictive accuracy and control overfitting [71] . This classifier has several parameters, e.g., criterion, maximum of features ( max_features ), minimum of the sample leaf ( min_sample_leaf ), and minimum of the sample split ( min_sample_split ).…”
Section: Resultsmentioning
confidence: 99%
“…It is crucial to assess the collective model with high accuracy to use the correct amount of randomness and the right approach. The notion of privacy in characterizing randomness analyze in the conventional privacy architecture, disclosure risk, and destruction metrics in data handling [11]; however, it describes in current designs [12].…”
Section: Related Workmentioning
confidence: 99%
“…Delete existing data and replacing it with new data is a logical technique to arbitrarily a collection of elements. Paper [11] looks into the choose-a-size group of arbitrariness algorithms. A choose-a-size arbitrariness operator is constructed for a fixed record size |r| = n and has three conditions: a arbitrariness level 0< ρ < 1 and a distribution function (d [0], d [1],.., d[n]) over the dataset {0, 1, .…”
Section: Dataset Arbitrarinessmentioning
confidence: 99%
“…It is the most accurate and computationally efficient algorithm. 55 It is different from the other tree-ensemble algorithms based on two reasons: first is that it divides the nodes by selecting the cut points fully at random basis and second one is that it takes the whole training data set (no subsets of the data set) to create more trees. Recently, ET ensemble is selected as the proposed scheme for ETD in SGs in Gunturi and Sarkar.…”
Section: Extra Treesmentioning
confidence: 99%