2023
DOI: 10.11591/ijece.v13i4.pp4683-4691
|View full text |Cite
|
Sign up to set email alerts
|

Bankruptcy prediction model using cost-sensitive extreme gradient boosting in the context of imbalanced datasets

Abstract: <span lang="EN-US">In the process of bankruptcy prediction models, a class imbalanced problem has occurred which limits the performance of the models. Most prior research addressed the problem by applying resampling methods such as the synthetic minority oversampling technique (SMOTE). However, resampling methods lead to other issues, e.g., increasing noisy data and training time during the process. To improve the bankruptcy prediction model, we propose cost-sensitive extreme gradient boosting (CS-XGB) t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
5
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(5 citation statements)
references
References 26 publications
0
5
0
Order By: Relevance
“…Similarly, feature selection and ensemble learning were additionally proposed in the economic sector. A light gradient boosting algorithm for risk analysis [50] and a cost sensitive-XGBoost approach for bankruptcy prediction [51] were developed with highly accurate results (around 95%) whilst [52] improved those outputs with an accuracy rate of 97%. On the other hand, principal component analysis (PCA) for feature selection [53] and Bayesian hyper-parameter optimization [54] were proposed in conjunction with XGBoost as optimal solutions for crowdfunding and credit worthiness prediction respectively.…”
Section: Resultsmentioning
confidence: 99%
“…Similarly, feature selection and ensemble learning were additionally proposed in the economic sector. A light gradient boosting algorithm for risk analysis [50] and a cost sensitive-XGBoost approach for bankruptcy prediction [51] were developed with highly accurate results (around 95%) whilst [52] improved those outputs with an accuracy rate of 97%. On the other hand, principal component analysis (PCA) for feature selection [53] and Bayesian hyper-parameter optimization [54] were proposed in conjunction with XGBoost as optimal solutions for crowdfunding and credit worthiness prediction respectively.…”
Section: Resultsmentioning
confidence: 99%
“…However, models like LR and SVM can struggle with unbalanced datasets. Ensemble methods, including Random Forest, AdaBoost, Gradient Boosting, XGBoost, and CatBoost, have improved predictive performance, with Random Forest and XGBoost, in particular, being known for their accuracy and robustness in handling diverse datasets [13]- [16].…”
Section: Related Workmentioning
confidence: 99%
“…To address this, class-balancing methods like under-sampling, oversampling (including SMOTE and its variants), and combined techniques are frequently used. Some studies also explore unbalanced datasets without prior class balancing [7], [8], [13], and [24]. The ongoing development in bankruptcy prediction reflects the continuous pursuit of more accurate, efficient, and reliable methods to anticipate financial distress and mitigate its effects.…”
Section: Related Workmentioning
confidence: 99%
See 2 more Smart Citations