2021
DOI: 10.1109/tr.2021.3118026
|View full text |Cite
|
Sign up to set email alerts
|

A Comparative Study of Class Rebalancing Methods for Security Bug Report Classification

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
35
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 72 publications
(35 citation statements)
references
References 43 publications
0
35
0
Order By: Relevance
“…The malicious samples (positive samples) come from the XSSed database and the tested payload (Payload) in the penetration test. Additionally, 75,428 pieces of standard data are obtained to ensure a balanced selection of samples ( Zheng et al, 2021 ; Cai et al, 2022 ). In the experiment, the training set and the test set are randomly selected from the samples at a ratio of 7:3.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…The malicious samples (positive samples) come from the XSSed database and the tested payload (Payload) in the penetration test. Additionally, 75,428 pieces of standard data are obtained to ensure a balanced selection of samples ( Zheng et al, 2021 ; Cai et al, 2022 ). In the experiment, the training set and the test set are randomly selected from the samples at a ratio of 7:3.…”
Section: Experiments and Resultsmentioning
confidence: 99%
“…The testing findings reveal that the suggested method's detection accuracy is 98.98 percent, indicating that it can successfully identify DDoS attack traffic in an SDN context [1]. Based on their analysis of the impact of class imbalance on SBR prediction, Zheng et al [32] found that it had a negative impact on prediction accuracy. A random forest classifier was used by Zhang et al [33] to train a Just-in-Time defect prediction model based on six open source projects.…”
Section: Wireless Communications and Mobile Computingmentioning
confidence: 96%
“…First, the GBDT algorithm is introduced to verify XGBoost results. e advantage of this algorithm is that it can deal with nonlinear data and is flexible [76][77][78]. In addition, it uses some robust loss functions, such as the Huber loss function and the Quantile loss function, which are very robust to outliers.…”
Section: Robust Testmentioning
confidence: 99%