2021
DOI: 10.1155/2021/4917016
|View full text |Cite
|
Sign up to set email alerts
|

A Novel Approach for Malicious URL Detection Based on the Joint Model

Abstract: The number of malicious websites is increasing yearly, and many companies and individuals worldwide have suffered losses. Therefore, the detection of malicious websites is a task that needs continuous development. In this study, a joint neural network algorithm model combining the attention mechanism, bidirectional independent recurrent neural network (Bi-IndRNN), and capsule network (CapsNet) is proposed. The word vector tool word2vec trains the character- and word-level uniform resource locator (URL) static … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
9
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 19 publications
(9 citation statements)
references
References 27 publications
(30 reference statements)
0
9
0
Order By: Relevance
“…Researchers use diverse datasets, including sources like PhishTank, Kaggle, CommonCrawl, GitHub, Phishstorm, Malcode, and DomainTools, to assess network detection and classification model efficacy, ensuring robustness and realworld relevance. In malicious website detection studies, features like HTML, JavaScript code, WHOIS host information, and web URL characteristics are manually extracted and incorporated into machine learning or heuristic systems for effective detection [5]. The training dataset for a classification model comprised 5 million URLs from Openphish, Alexa whitelists, and internal FireEye sources, maintaining a balanced 60-40 split between benign and malicious URLs [8].…”
Section: Datasets Usedmentioning
confidence: 99%
See 1 more Smart Citation
“…Researchers use diverse datasets, including sources like PhishTank, Kaggle, CommonCrawl, GitHub, Phishstorm, Malcode, and DomainTools, to assess network detection and classification model efficacy, ensuring robustness and realworld relevance. In malicious website detection studies, features like HTML, JavaScript code, WHOIS host information, and web URL characteristics are manually extracted and incorporated into machine learning or heuristic systems for effective detection [5]. The training dataset for a classification model comprised 5 million URLs from Openphish, Alexa whitelists, and internal FireEye sources, maintaining a balanced 60-40 split between benign and malicious URLs [8].…”
Section: Datasets Usedmentioning
confidence: 99%
“…Since they are usually disguised to resemble trustworthy websites, they pose a threat to unwary users. According to a survey by Kaspersky [5], 173 million dangerous URLs were detected by web security software in 2020. Additionally, the report also indicated that 66.07% of the malicious URLs were 20 of the most recent harmful apps.…”
Section: Introductionmentioning
confidence: 99%
“…They used two datasets: the first dataset (DS1) was compiled from PhishTank[14] and Alexa [16]. The second dataset (DS2) was a bench-mark dataset used by Sahingoz et al [130] from PhishTank[14] and Yandex [137]. They extracted lexical features (character embedding features) using the CNN model and classified multilevel features using RF classifiers.…”
Section: ) Lexical Studiesmentioning
confidence: 99%
“…In general, malicious URLs can be identified with the help of Machine Learning (ML) technique through two steps as detailed herewith. At first, a suitable feature indication is obtained from the URL, and secondly, based on the feature identified, ML-related prediction methods are provided training to find out the malicious URLs [7,8]. The first step discussed above i.e., attaining the feature indication in which fruitful information regarding the URL is saved in a vector so that the ML methods can be implied to it.…”
Section: Introductionmentioning
confidence: 99%