2022
DOI: 10.5755/j01.itc.51.2.30380
|View full text |Cite
|
Sign up to set email alerts
|

Improved Detection of Malicious Domain Names Using Gradient Boosted Machines and Feature Engineering

Abstract: Malicious domain names have been commonly used in recent years to launch different cyber-attacks. There are a large number of malicious domains that are registered every day and some of which are only active for brief periods of time. Therefore, the automated malicious domain names detection is needed to provide security for individuals and organisations. As new technologies continue to emerge, the detection of malicious domain names remains a challenging task. In this study, we propose a model to effectively … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
3
1
1

Relationship

1
4

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 31 publications
0
3
0
Order By: Relevance
“…The first work for a browser-based application utilizing only machine learning is MADMAX [3], including feature selection. In recent years, MADDONA [4] was proposed as an extension of MADMAX to optimize feature selection and neural network architectures further in accordance with the latest feature engineering [30]. We combine our dataset with MADMAX and then show that its resultant performance outperforms MAD-DONA.…”
Section: Further Related Workmentioning
confidence: 98%
“…The first work for a browser-based application utilizing only machine learning is MADMAX [3], including feature selection. In recent years, MADDONA [4] was proposed as an extension of MADMAX to optimize feature selection and neural network architectures further in accordance with the latest feature engineering [30]. We combine our dataset with MADMAX and then show that its resultant performance outperforms MAD-DONA.…”
Section: Further Related Workmentioning
confidence: 98%
“…To choose the most influential feature, we utilize the automatic feature engineering technique to improve the model performance employing the power of the most relevant features [23]. Automatic feature engineering is the process of automatically selecting or building new features from existing features [24], which trade-off between feature complexity and model performance.…”
Section: E Automatic Feature Engineeringmentioning
confidence: 99%
“…Entropy measures the uncertainty of the data. From a different perspective, entropy measures how difficult it is to guess the label of a random sample from a dataset, where low entropy indicates that the data labels are quite uniform, and high entropy indicates that the labels are in confusion [21]. Information gain computes the difference between the entropy before and after a split and specifies class element impurity.…”
Section: Feature Importancementioning
confidence: 99%