2023
DOI: 10.3390/ijgi12050197
|View full text |Cite
|
Sign up to set email alerts
|

Evaluating Landslide Susceptibility Using Sampling Methodology and Multiple Machine Learning Models

Abstract: Landslide susceptibility assessment (LSA) based on machine learning methods has been widely used in landslide geological hazard management and research. However, the problem of sample imbalance in landslide susceptibility assessment, where landslide samples tend to be much smaller than non-landslide samples, is often overlooked. This problem is often one of the important factors affecting the performance of landslide susceptibility models. In this paper, we take the Wanzhou district of Chongqing city as an exa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
9

Relationship

0
9

Authors

Journals

citations
Cited by 16 publications
(6 citation statements)
references
References 53 publications
0
6
0
Order By: Relevance
“…In this research, the aim was not to control all variables but to maximize the available resources; therefore, the uncertainty regarding the number and type of variables was minimized by conducting a comprehensive analysis of the topographic variables in the models, including correlation, multicollinearity, and dimensional reduction using PCA. This approach allowed us to exclude variable correlations, reduce noise, and mitigate the risk of overfitting, thus improving the accuracy of the models [45,81]. Additionally, the WoE analysis was employed to identify causal relationships between instability factors and the distribution of MM.…”
Section: Discussionmentioning
confidence: 99%
“…In this research, the aim was not to control all variables but to maximize the available resources; therefore, the uncertainty regarding the number and type of variables was minimized by conducting a comprehensive analysis of the topographic variables in the models, including correlation, multicollinearity, and dimensional reduction using PCA. This approach allowed us to exclude variable correlations, reduce noise, and mitigate the risk of overfitting, thus improving the accuracy of the models [45,81]. Additionally, the WoE analysis was employed to identify causal relationships between instability factors and the distribution of MM.…”
Section: Discussionmentioning
confidence: 99%
“…Comparing the variables selected by the PCA with those in SET-3 shows substantial agreement. However, when comparing with those selected in [68], it is found that both algorithms use similar variables in their predictions as follows: lithology, earthquake intensity parameters, and curvature. These matches in variable selection highlight the importance of these factors in predicting co-seismic landslides.…”
Section: Validation and Comparison With Physical Methodologiesmentioning
confidence: 99%
“…Its objective is to obtain a new set of uncorrelated variables called "principal components", which capture the most information in terms of the variance of the original data. This technique is widely used in various fields [68] to perform tasks such as dimensionality reduction, data visualization, pattern detection, and exploration of the underlying structure in a set of variables.…”
Section: Model Evaluationmentioning
confidence: 99%
“…Looking at the previous studies using ML algorithms, it is seen that some studies produced LSMs with a single ML algorithm [14][15][16][17], while some studies produced LSMs using multiple ML algorithms together [18][19][20] and compared their performances. In these studies, algorithms such as Support Vector Machines (SVM) [21,22], K-Nearest Neighbor (KNN) [23,24], Naïve Bayes (NB) [25,26], Artificial Neural Network (ANN) [27,28], Multilayer Perceptron (MLP) [7,29], Classification and Regression Tree (CART) [30,31], Random Forest (RF) [16,32], Adaptive Boosting (AdaBoost) [33,34], Gradient Boosting Machine (GBM) [28,35], Light Gradient Boosting Machine (LightGBM) [36,37], Natural Gradient Boosting (NGBoost) [3], Extreme Gradient Boosting (XGBoost) [17] and categorical boosting (CatBoost) [36,38] are frequently used. Algorithms such as RF, GBM, LightGBM, AdaBoost, NGBoost, CatBoost, and XGBoost that use ensemble methods such as bagging, stacking, or boosting are called tree-based ensemble learning algorithms [36,39].…”
Section: Introductionmentioning
confidence: 99%