2017
DOI: 10.1109/access.2017.2738069
|View full text |Cite
|
Sign up to set email alerts
|

An Ensemble Random Forest Algorithm for Insurance Big Data Analysis

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
93
0
7

Year Published

2018
2018
2023
2023

Publication Types

Select...
8

Relationship

0
8

Authors

Journals

citations
Cited by 264 publications
(101 citation statements)
references
References 16 publications
1
93
0
7
Order By: Relevance
“…, x l (172)] and y l ∈ {1, 2, 3, 4, 5, 6, 7, 8} denote the input 172 features and the output label of sample l, the general idea of random forest can be described as, (1) Randomly select N samples with replacement from the original dataset, and obtain N subsamples for constructing each tree. As an ensemble model, random forest model fits the input data in a shorter time as each decision tree is independent, making parallel computing and modeling possible [31,32]. We also test the performance of the other five machine learning methods, including Fine Tree, Radial Basis Function kernel Support Vector Machine (RBF SVM), Weighted K-Nearest Neighbors (KNN), Linear Discriminant and Subspace KNN [33,34].…”
Section: Random Forest For Image Classificationmentioning
confidence: 99%
See 1 more Smart Citation
“…, x l (172)] and y l ∈ {1, 2, 3, 4, 5, 6, 7, 8} denote the input 172 features and the output label of sample l, the general idea of random forest can be described as, (1) Randomly select N samples with replacement from the original dataset, and obtain N subsamples for constructing each tree. As an ensemble model, random forest model fits the input data in a shorter time as each decision tree is independent, making parallel computing and modeling possible [31,32]. We also test the performance of the other five machine learning methods, including Fine Tree, Radial Basis Function kernel Support Vector Machine (RBF SVM), Weighted K-Nearest Neighbors (KNN), Linear Discriminant and Subspace KNN [33,34].…”
Section: Random Forest For Image Classificationmentioning
confidence: 99%
“…As an ensemble model, random forest model fits the input data in a shorter time as each decision tree is independent, making parallel computing and modeling possible [31,32]. We also test the performance of the other five machine learning methods, including Fine Tree, Radial Basis Function kernel Support Vector Machine (RBF SVM), Weighted K-Nearest Neighbors (KNN), Linear Discriminant and Subspace KNN [33,34].…”
Section: Random Forest For Image Classificationmentioning
confidence: 99%
“…An Ensemble Random Forest Algorithm for Insurance Big Data Analysis [4] by W. Lin proposed an A group arbitrary woods calculation which utilized the parallel registering capacity and memory-store system improved by Spark. The proposed technique The group arbitrary backwoods calculation outflanked SVM and other characterization calculations in both execution and exactness utilizing the imbalanced information, and it is helpful for improving the precision of item showcasing contrast with the customary fake methodology.…”
Section: A Enormous Data Classificationmentioning
confidence: 99%
“…Enormous information innovation was generally connected. In the scholastic network, the regarded diaries "Nature" and "Science" have separately propelled huge information issues named "Enormous Data" and "Manage Data", which examine an assortment of issues experienced in huge information innovation from the Internet innovation, financial matters, supercomputing, organic sciences, drug and numerous different viewpoints [4]. The monstrous volume of data assembled by contemporary frameworks wound up inescapable, the same number of research exercises require gathering progressively immense measures of information [1].…”
Section: Introductionmentioning
confidence: 99%
“…Second, random forest is suitable for handling large data due to its parallelization [28]. It has been combined with the Spark [28], heuristic bootstrap sampling method [29], kernel principal component analysis [30], and other technologies to perform fault diagnosis and regression tasks [31,32]. Owing to the improvement of the forecasting accuracy for highdimensional and large-scale wind turbine data, we propose an optimized random forest method which consists of a dimension reduction procedure and the weighted voting process for the short-term WPF.…”
Section: Mathematical Problems In Engineeringmentioning
confidence: 99%