2020
DOI: 10.1016/j.compbiomed.2020.103899
|View full text |Cite
|
Sign up to set email alerts
|

Improving protein-protein interactions prediction accuracy using XGBoost feature selection and stacked ensemble classifier

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
73
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
3

Relationship

0
9

Authors

Journals

citations
Cited by 172 publications
(74 citation statements)
references
References 70 publications
1
73
0
Order By: Relevance
“…We adopted XGBoost and the stacking ensemble technique, which has been widely used to achieve better performance in many Kaggle competitions and research. 27,28 Recently, the combination of single XGBoost models using a stacking ensemble technique showed better performance than other machine learning algorithms did, which is consistent with our observations. 29 In our experience, the proposed ensemble technique showed superior performance to that of other state-of-the-art machine learning methods through trial and error.…”
Section: Discussionsupporting
confidence: 90%
“…We adopted XGBoost and the stacking ensemble technique, which has been widely used to achieve better performance in many Kaggle competitions and research. 27,28 Recently, the combination of single XGBoost models using a stacking ensemble technique showed better performance than other machine learning algorithms did, which is consistent with our observations. 29 In our experience, the proposed ensemble technique showed superior performance to that of other state-of-the-art machine learning methods through trial and error.…”
Section: Discussionsupporting
confidence: 90%
“…As the publicly available virus-host PPI data increased, the emphasis on this subject has recently been shifted to machine-learning-based computational techniques to identify virus-host PPIs. PPI prediction tools have been developed based on different machine-learning models such as support vector machines (SVM) (Shen et al, 2007;Cui et al, 2012;Eid et al, 2016), random forest (RF) (Yang et al, 2020) and gradient boosting machine (XGBoost) (Basit et al, 2018;Chen et al, 2020).…”
Section: Machine Learning-based Ppi Prediction Toolsmentioning
confidence: 99%
“…To determine important features, the gain is used for the optimal node split during training in the tree-based classifiers such as RF, XGB, and GBC by Eq. (12) [43]:…”
Section: ) Feature Importance and Cumulative Feature Importancementioning
confidence: 99%