0% found this document useful (0 votes)
139 views

Fake URL Detection Using Machine Learning and Deep Learning

The risk of network information insecurity is growing rapidly in number and level of risk is very high.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
139 views

Fake URL Detection Using Machine Learning and Deep Learning

The risk of network information insecurity is growing rapidly in number and level of risk is very high.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology

ISSN No:-2456-2165

Fake URL Detection Using Machine


Learning and Deep Learning
Vedav K S Koushik Nayak U
Department of Computer Science, Department of Computer Science,
Dayananda Sagar Collage of Engineering Dayananda Sagar Collage of Engineering
Bangalore, India Bangalore, India

A Mukesh Karthik V
Department of Computer Science, Department of Computer Science,
Dayananda Sagar Collage of Engineering Dayananda Sagar Collage of Engineering
Bangalore, India Bangalore, India

Soumya Patil
Department of Computer Science,
Dayananda Sagar Collage of Engineering
Bangalore, India

Abstract:- The risk of network information insecurity is There are several scientific studies showing many malicious
growing rapidly in number and level of risk is very high. URL detection methods based on machine learning and deep
The methods mostly used by hackers today is to attack learning techniques. This document proposes a method to
whole system and exploit human vulnerabilities. These detect spoofed URLs using machine learning techniques
techniques include social engineering, phishing, based on proposed URL behaviours and attributes. In
pharming, etc. One of the steps in conducting these addition, big data technology is also leveraged to enhance the
attacks is to deceive users with fake Uniform Resource ability to detect malicious URLs based on their anomalous
Locators (URLs). As a result, fake URL detection is of behaviour. In short, the proposed detection system consists of
great interest nowadays. There have been several novel features and behaviours of URLs, machine learning
scientific studies showing a number of methods to detect algorithms, and big data techniques. Experimental results
malicious URLs based on machine learning and deep show that the proposed URL attributes and behaviours help
learning techniques. In this paper, we propose a Fake significantly improve detection of malicious URLs. This
URL detection method using machine learning techniques indicates that the proposed system can be viewed as a
based on our proposed URL behaviours and attributes. streamlined and easy-to-use malicious URL detection
Moreover, bigdata technology is also exploited to improve solution. URLs (Uniform Resource Locators) are used to
the capability of detection malicious URLs based on refer to resources on the Internet. [1] presents the properties
abnormal behaviours. In short, the proposed detection and two basic components of a URL as a protocol identifier,
system consists of a new set of URLs features and which indicates the protocol to use, and a resource name,
behaviours, a machine learning algorithm, and a bigdata which indicates the IP address or domain name where the
technology. The experimental results show that the resource is located. You can see that each URL has a specific
proposed URL attributes and behaviour can help improve structure and format. This can be suggested and identified
the ability to detect malicious URL significantly. This is when an attacker attempts to change one or more of her details
suggested that the proposed system may be considered as in her URL. Malicious URLs are known as links that harm
an optimized and friendly used solution for malicious users. These URLs are resources or pages that allow attackers
URL detection. to execute code on your computer, redirect you to unwanted,
malicious, or other phishing sites, or download malware.
Keywords:- URL; Malicious URL Detection; Phishing; redirect the user to Malicious URLs can be found in
Machine Learning everything from how files are downloaded to how movies are
downloaded, drive-by downloads, phishing, spamming,
I. INTRODUCTION tampering, and more.

The risk of network information becoming unstable is A. Clayton Johnson, Bishal Khadka, Ram B. Basnet
growing rapidly, and the level of risk is very high. The Organizations face significant threats from emails with
primary method used by hackers today is to attack entire Uniform Resource Locators (URLs), which may compromise
systems and exploit human vulnerabilities. These techniques network security and user credentials through spear-phishing
include social engineering, phishing, pharming, and more. and other common phishing techniques. campaigns to their
One of the steps in carrying out these attacks is to trick users staff. The identification and classification of harmful URLs is
with a fake URL (Uniform Resource Locator). That's why a crucial practical application to a scientific challenge. An
there's a lot of interest in detecting fake URLs these days. organisation can safeguard itself by filtering incoming emails

IJISRT22DEC1128 www.ijisrt.com 1198


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
and the websites that its employees are accessing with the trustworthy websites rated by Alexa could be corrupted or
help of the right machine learning model, depending on the fake URLs that have been defaced. In this study, we
maliciousness of URLs found in emails and web pages. In this investigate a quick method for identifying and classifying
work, we compare the performance of conventional machine dangerous URLs using their style of assault. We demonstrate
learning methods, such as Random Forest, CART, the value and effectiveness of lexical analysis.
Comparing kNN against well-known deep learning
framework models as Fast.ai and Keras-TensorFlow spans E. Tie Li a, Gang Kou b, Yi Peng a
CPU, GPU, and TPU architectures. Using the ISCX-URL- Traditional classifiers are challenged in the detection of
2016 dataset, which is accessible to the general public. We dangerous URLs due to the enormous volume of data. The
display the models' results over binary. relationships between attributes are intricate, and patterns are
evolving over time. Feature In order to solve these
B. Vinayakumar R, Sriram S, Soman KP, and Mamoun difficulties, engineering is crucial. To more accurately depict
Alazab the underlying This research provided a method to resolve the
Malicious Uniform Resource Locator (URL), also issue and enhance the capabilities of classifiers in identifying
referred to as a malicious website is the main platform for malicious URLs. a method of spatial translation that
hosting unsolicited content such spam, malicious ads, combines linear and non-linear techniques. To change
phishing, and drive-by downloading, escapades, to mention a something linearly, A two-stage distant learning
few. Identifying harmful actors is essential timely URLs. methodology was created. The first step was singular value
Blacklisting, previously employed in studies techniques using decomposition an orthogonal space was created, and a linear
regular expression and signature matching. These techniques programming technique was employed to solve an ideal
are absolutely unsuccessful at detecting variations of a distance measurement. Nyström method for kernel nonlinear
previously discovered URL that is dangerous or a completely transformation was introduced using the updated distance
new URL. This by suggesting the machine learning-based metric for its radial basis function, the merits of this approach
solution, the problem can be reduced, solution. Such a were approximated. There were 33,1622 URLs with 62
solution necessitates a thorough investigation in Security characteristics gathered to verify the suggested feature
artefact feature engineering and feature representation enter engineering techniques. The outcomes demonstrated that the
something like URLs. Additionally, resources for feature suggested approaches dramatically increased the
engineering and feature representation must be continuously effectiveness and performance of several classifiers, such the
improved to handle variations on current URLs or completely k-Nearest Neighbor classifier. Neural networks, support
new URLs. vector machines, and neighbours. The percentage of
malicious URLs that were found was the rate of the linear
C. Shantanu, Janet B, Joshua Arul Kumar R Support Vector Machine was enhanced from 68% to 86%,
One of the most frequent cybersecurity threats is a and k-Nearest Neighbor was, The rate of Multi-Layer
malicious universal resource locator (URL), or malicious Perceptron also climbed, from 63% to 82%, while the rate of
websites, threats. They lure naïve visitors to sign up by both decreased from 58% to 81%. We additionally created a
hosting gratuitous content (such as spam, malware, webpage to showcase a malicious URLs detection system that
inappropriate adverts, and spoofing) and victims of scams makes use of the techniques recommended in this document.
(money loss, exposure of personal details, installation of
malware, extortion, a false online store, and unexpected F. Zhiqiang Wang, Xiaorui Ren, Shuhao Li, Bingyan Wang,
reward, etc.) and annually result in billions of rupees in Jianyi Zhang , and Tao Yang
losses. The trip Email, ads, and web searches may all bring Network security is vulnerable to a variety of dangers as
traffic to these websites or a website's ties to another. Each Internet technology advances. Attackers, in particular, can
time, the user needs to click on the rogue URL. The increase disseminate harmful universal resource locators (URL) are
in phishing and spam incidents and malware has created a used to conduct assaults like spam and phishing. study on
pressing need for a trustworthy remedy which can categorise harmful URLs For the purpose of fighting against these
and recognise dangerous URLs. Traditional categorization attacks, detection is important. The current research does still
methods include regular expression, blacklisting, and have some issues, though. For instance, it is difficult to
signature matching methods are difficult effectively isolate harmful traits. Some of the current
detection techniques are simple for attackers to sidestep. We
D. Mohammad Saiful Islam, Mamun, Mohammad Ahmad To address these issues, create a dynamic convolutional
Rathore, Arash Habibi Lashkari, Natalia Stakhanova, neural network (DCNN)-based malicious URL detection
and Ali A. Ghorbani model to the initial multilayer convolution network, a new
The Internet has long since evolved into a significant folding layer is added. It substitutes the k-max-pooling layer
hub for online criminal activity. In this domain, URLs are the for the pooling layer. The depth of the feature mapping in the
primary method of communication to the security community middle of the dynamic convolution algorithm. Additionally,
concentrated its efforts on creating methods mostly for the settings of the pooling layer are dynamically changed in
blacklisting harmful URLs. whilst thriving. This method only accordance with the length of the URL input and the depth of
partially succeeds in shielding users from known harmful the current convolution layer, which is advantageous for
domains part of the issue is resolved. The fresh dangerous obtaining more detailed information across a larger spectrum.
URLs that appeared everywhere online in large numbers This study examines I suggest a novel embedding technique
frequently have a head start in this race. Besides even that makes use of word embedding based on character

IJISRT22DEC1128 www.ijisrt.com 1199


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
embedding to learn the vector a picture of a URL. We carry methodology can be used to construct a real-time website
out two sets of comparative experiments in the interim. First, defacement monitoring system because it doesn't require a lot
we compare three things, experiments that use several of computational power.
embedding techniques and the same network structure. The
outcomes demonstrate that word embedding based on J. Ashit Kumar Dutta
character embedding, greater precision can be attained. Modernizations in Internet and cloud technologies have
produced a large a growth in consumer online purchases and
G. Cheng Cao and James Caverlee other forms of electronic trading. The damage caused by this
This study tackles the problem of identifying spam increase, which allows illegal access to users' private
URLs protecting users from links is a crucial duty in social information, resources of a business. One of the well-known
media connected with malware, phishing, and other suspect, attacks that deceives people into accessing to spread
low-quality content. Rather than relying on content or filters dangerous content and collect their data. In terms of unified
with historical blacklists examination of the Web URLs' website interface the resource location (URL), the majority of
landing pages, we look at the behavioural factors related to phishing websites mimic legitimate websites exactly. There
both the URL's poster and its clicker. The fundamental are many methods for spotting phishing websites, including
assumption is that it might be harder to detect these blacklists, heuristics, etc. suggested. However, there is an
behavioural signs than conventional signals, manipulate. exponential increase in cybercrime due to ineffective security
Specifically, we recommend and assess fifteen features that systems the number of casualties has increased. The
are click- and posting-based. After much experimentation. unpredictable and anonymous foundation of the Internet users
We find that this purely behavioural approach can achieve are more susceptible to phishing scams. Existing research
high results in our evaluation area-under-the-curve (0.92), demonstrates that the phishing detection system's
recall (0.86), and precision (0.86) all point to the possibility performance is constrained. A demand exists for an
of robust behavior-based spam. intelligent a method to shield users against cyberattacks. The
author of this study suggested a URL based on machine
H. Christophe Chong, Daniel Liu, and Wonhong Lee learning methods, detecting technology. a neural network
The adoption of smartphones and other mobile devices with recurrence a technique is used to identify phishing
for both personal and professional purposes have increased URLs. Researchers assessed the suggested strategy contains
web vulnerability used professionally. In this research, we 5800 trustworthy websites and 7900 harmful ones,
present a machine learning approach to detect dangerous respectively. The result of the experiments demonstrates that
URLs by combining of payload size, JavaScript source the performance of the suggested strategy outperforms that of
features, and URL lexical features. We employ a polynomial current approaches in malware detection in URLs.
kernel SVM with get an F1 score of 0.74 and an accuracy of
0.81. K. Frank Vanhoenshoven, Gonzalo Napoles , Rafael Falcon,
Koen Vanhoof and Mario Koppen
I. Xuan Dau Hoang and Ngoc Tuong Nguyen The World Wide Web accommodates a variety of
For a long time, defacement attacks have been regarded criminal acts include financial fraud and e-commerce with
as one of the main hazards to websites and services spam advertisements fraud and the spread of viruses.
applications made by businesses, enterprises, and Although the particular reasons behind these plans may vary,
governmental bodies. Attacks by vandals may result in they all share one thing in common lies in the unknowing
significant repercussions for website owners, including an consumers that frequent their websites. Those trips email,
instant suspension of website activities and reputational harm online search engine results, or links from other websites a
to the owner, which could lead to significant financial losses. website. However, the user is always forced to take some
several options for monitoring and detecting website example as selecting a suitable Uniform Resource Locator by
defacement threats, research and deployment such as those clicking(URL). In order to identify these fraudulent sites, the
relying on complex DOM tree analysis, checksum web security Blacklisting services have been established by
comparison, and diff comparison algorithms. However, some the community. Such blacklists are subsequently created
solutions are only applicable to static websites, while others using a variety of methods, such as web crawlers, honeypots,
call for substantial processing power. The hybrid defacement and manual reporting paired with strategies for site analysis.
detection model proposed in this paper is based on the mix of In this article, we discuss how to identify rogue URLs as a
signature-based detection with machine learning-based binary classification issue and research the results a number
detection. The device A detection profile is first created by of well-known classifiers, including Naive Bayes, Support
learning-based detection using training data from both normal Vector k-Nearest Neighbors, Random Forest, Decision Trees,
and hacked websites Afterward, it makes use of the profile to Multi-Layer Perceptrons, and Vector Machines.
categorise tracked web pages as either normal or attacked. Additionally, we adopted 2.4 million URLs (instances) in a
The machine learning-based element may successfully public dataset 3,2,000,000 features The mathematical
identify tampering for both static and dynamic pages and calculations have demonstrated that. The majority of
pages. However, the signature-based detection is employed to categorization techniques yield respectable prediction rates
increase the performance of the model in analysing typical without using either sophisticated feature selection methods
defacements. Numerous experiments demonstrate that Our or the assistance of a subject matter expert. Specifically,
model generates a false positive rate of roughly 0.27% and an Random The highest levels of accuracy are achieved by forest
overall accuracy of more than 99.26%. Additionally, our and multi-layer perceptrons.

IJISRT22DEC1128 www.ijisrt.com 1200


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
L. Tiefeng Wu, Miao Wang , Yunfang Xi and Zhichao Zhao This study Should be pertinent to cybersecurity experts,
As Internet technology has advanced quickly, a large academic researchers because it might serve as a foundation
number of dangerous URLs have emerged, posing numerous for real-life detecting technologies or additional research.
security hazards. Detecting dangerous URLs quickly has
emerged as a crucial component of cyberattack defence. Deep O. Immadisetti Naga Venkata Durga Naveen, Manamohana
learning techniques result in new improvements in K, Rohit Verma
identifying dangerous web pages. This paper suggests a The earliest form of URL (Uniform Resource Locator)
harmful URL a detection technique based on an attention as a web address. Nevertheless, some URLs could be utilised
mechanism and a bidirectional gated recurrent unit (BiGRU). to house unwelcome content that may result in online
The BiGRU model is the foundation of the technique. A assaults. Malicious URLs are what we refer to as these. The
regularisation operation called a dropout mechanism is an end user system's incapacity to find and get rid of harmful
attention mechanism is also introduced to the input layer to URLs might leave a trusting person exposed condition.
prevent the model from overfitting to the middle layer to Additionally, using malicious URLs could result in
improve the ability to learn URL features. adversary's unauthorised access to the user data. The primary
reason, They offer an attack for malicious URL detection
M. S. Markkandeyan, C. Anitha surface to the opponent. It is crucial to stop these actions by
The World Wide Web's use and benefits have permeated using several novel methods. Numerous literary works have
every aspect of daily life for people, including transferring been techniques for blocking out dangerous URLs. a few of
information and spreading knowledge so quickly and readily them are Heuristic Classification, Black-Listing Therefore,
in time. search for theft and Phishing is one of the two forms these traditional techniques are ineffective properly handle
of cybercrime when hackers and malevolent users steal the constantly changing technologies and webaccess
personal information on the actual, legitimate users who are methods. Furthermore, these methods are ineffective in
using it to make unlawful financial gains. malignant URLs identifying contemporary URLs like short URLs, black web
host a variety of enticing incidents such phishing, spam, URLs. In this article, we suggest a fresh classification
drive-by vulnerabilities, and so forth and duping the trusting approach to solve the difficulties that the conventional
people into become the target of such frauds by experiencing processes in malware detection in URLs. The suggested
financial loss, data loss, and malware installation, etc causing categorization scheme is constructed using advanced
the victims to sustain catastrophic losses amounting to machine learning techniques that not only focuses on the
billions of dollars each year. Historically, this type of fraud URL's syntactic structure as well as the lexical and semantic
has been found utilising the blacklists, which are not content of these dynamic URLs. The proposed strategy is
exhaustive and additionally lacking the capacity to recognise anticipated to perform better than the already in use methods.
newly created infamous and dangerous URLs. Consequently,
to identify given these horrible crimes, it is urgent to P. malak aljabri, hanan S. Altamimi , shahd A.Albelali,
implement a system that is fool proof and has wider maimunah al-harbi,haya T. Alhuraib, najd K. Alotaibi,
implications with its speed and accuracy to identify the source amal A. Alahmadi,fahd alhaidari
and advocate of such malicious contents. The spontaneous, The digital world has grown tremendously in recent
One of such open systems is the Convolution Neural Network years, especially on the Internet, which is essential because so
(CNN) model system the authors propose hindered by the many of our activities are now carried out online. Due to the
inability to re-learn. assailants' creative the likelihood of a cyberattack is growing
quickly. The malevolent attack is one of the most important
N. Adebayo Oshingbesan Kagame Richard Aime Munezero ones. URL designed to deceive unskilled end users into
Courage O Ekoh providing unrequested information, leading to putting the
A typical method for identifying fraudulent websites user's system at risk and resulting in annual losses of billions
using blacklists, which are not all-inclusive, as a strategy they of dollars. Consequently, obtaining online presence is getting
remain specific to themselves and cannot spread to new increasingly important. In this essay, we present a thorough
malicious sites the identification of newly discovered overview of the literature, highlighting the primary methods
dangerous websites automatically will assist in lowering this for identifying fraudulent URLs are based on machine
form's susceptibility to attack of assault. In this research, we learning models, considering the limits of the datasets
looked at eleven machine learning algorithms to categorise employed, the feature types, detection technologies, and
dangerous websites using lexical data features and literature Additionally, we stress the guidelines due to the
comprehend how they apply to different datasets. We trained, dearth of studies on the detection of harmful Arabic websites
verified, and tested these models specifically on subsequently studies in this situation. Last but not least, following our
performed a cross-datasets analysis using various sets of examination of the chosen papers, we give obstacles that
datasets analysis. According to our investigation, K-Nearest, could lower the effectiveness of malicious URL detectors, as
The only model that consistently delivers strong results is well as potential solutions.
Neighbour. Additional models, including Random Forest,
Support vector machines, decision trees, and logistic Q. Fuqiang Yu
regression Additionally, machines consistently surpass a How can we combat Internet viruses is a difficult and
model of identifying each link across all metrics as harmful, pressing issue make sure search engines are secure. a search
datasets. In addition, we discovered no proof that any engine's security component based on the development of the
segment of lexical features are cross-model or cross-dataset. V2.0 investigation of the content-based image search engine

IJISRT22DEC1128 www.ijisrt.com 1201


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
system. A negative method for detecting URLs (Uniform with machine learning-based detection. The device A
Resource Locators) based on the Boyer-Moore pattern. It is detection profile is first created by learning-based detection
suggested to match. These are the primary research findings using training data from both normal and hacked websites
and contents: Web image searches may download malicious Afterward, it makes use of the profile to categorise tracked
URLs, which could lead to users suffer needless losses. web pages as either normal or attacked. The machine
Consequently, the BM-based dangerous URL detection learning-based element may successfully identify tampering
algorithm Matching patterns is suggested. This approach for both static and dynamic pages and pages. However, the
enables the virus to match the URL source code identifying signature-based detection is employed to increase the
features in the database to determine whether the URL is performance of the model in analysing typical defacements.
secure or not. Web image 203 dangerous URLs are found by Numerous experiments demonstrate that Our model has a
search using this technique. Kaspersky scanning allows us to false positive rate of roughly 0.27% and an overall accuracy
189 URLs have been determined to be malicious, with a 6.9% of more than 99.26%.
mistake rate and an accurate percent is 93.1%. The testing
outcomes demonstrate that the algorithm for detecting T. Trong Hung Nguyen1,Xuan Dau Hoang2,Duc Dung
dangerous URLs safe URLs for picture search engines on the Nguyen
web. Recently, defacement and general web attacks,
particularly those targeting websites and web applications,
R. Kevin Borgolte, Christopher Kruegel, Giovanni Vigna one of the top security risks to many businesses, and
Through the loss of sales, website vandalism and companies that offer web-based services. a tampering attack
defacements can cause the website owner significant harm could have a serious impact on the owner's website, including
due to a drop in reputation or legal repercussions. Prior efforts as well as immediate halting of website operations and harm
to detect website defacements have focused on identifying to the owner's reputation, which could result in significant
unapproved changes to the web server, for example, by file- financial losses. Several methods, metrics, and instruments
based integrity or host-based intrusion detection systems for website defacement monitoring and detection have been
checks. However, the majority of earlier methods are unable research, development, and practical application. Even so,
to recognise the most common defacement methods currently some Only static web pages can be used for measures and
in use. Currently: DNS hijacking and attacks including code approaches. other people can use dynamic web sites, but they
and/or data injection. This is due to the fact that these attacks demand a lot of computing power. The other problems with
don't actually change the Rather than changing the website's existent ideas have a high false positive rate and a poor
setup or source code, they add fresh content or send visitors detection rate alarming rate because many crucial
to another site. This essay tackles the issue of defacement components of websites, like images and embedded code are
detection from an alternative perspective: we employ not processed. To be able to In order to resolve these
computer vision ways to tell whether a website has been problems, this research suggests a combination model for
vandalised, similarly how a human analyst determines website defacement, based on BiLSTM and Efficient Net
whether a webpage has been altered when using a web detection. The suggested approach processes two types of
browser to view it. We present MEERKAT a system that can web pages: vital elements, such as the page's content and the
detect defacements without previous knowledge of the text screen captures The combination model may be
website's structure or content, but just the URL. When a successful it can achieve excellent detection rates with
defacement is discovered, the system alerts the website's dynamic web pages precision and a low number of false
owner that his site has been vandalised, who then be able to alarms. experimentation with Over 96,000 online pages in a
act appropriately. To identify tampering, MEERKAT dataset show that the suggested model performs better than
automatically picks out significant characteristics from most other models on most metrics. The F1-score, false
screenshots of websites that have been altered using current positive rate, and total accuracy of the model are 97.49%,
machine learning innovations like stacked autoencoders 96.87%
along with deep neural networks and computer vision.
U. Kevin Borgolte
S. Xuan Dau Hoang and Ngoc Tuong Nguyen It is simple to communicate and engage with people
Defacement assaults are often regarded as one of the top throughout the world due to the broad availability of web-
dangers to websites and services applications made by based services and Internet access. Unfortunately, attackers
businesses, enterprises, and governmental bodies. Attacks by frequently target the software and protocols used to
vandals may result in significant repercussions for website implement the functionality of these services. In turn, a
owners, including an instant suspension of website activities perpetrator can use them to She would compromise, seize
and reputational harm to the owner, which could lead to control, and misuse the services for her own evil ends. This
significant financial losses. several options for monitoring dissertation includes We develop techniques and algorithms
and detecting website defacement threats, research and to identify and mitigate such attacks in an effort to better
deployment such as those relying on complex DOM tree understand them. Using extensive datasets, we examine
analysis, checksum comparison, and diff comparison methods to stop them. The system Meerkat, which can
algorithms. However, some solutions are only applicable to identify website defacements as a visible sign of a
static websites, while others call for substantial processing compromised website, is described first. They have the
power. The hybrid defacement detection model proposed in potential to do the owners of the websites great harm either
this paper is based on the mix of signature-based detection as a result of decreased sales, diminished reputation, or legal

IJISRT22DEC1128 www.ijisrt.com 1202


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
repercussions. Meerkat demands without prior knowledge of results in varying degrees of cyberwarfare. The development
the websites' structure or content, merely having access to the of novel activities, such as social Attackers now have more
uniform resource Identifier (URI) where you can find them. options to find weaknesses thanks to networking, the
Meerkat intentionally imitates a human analyst's style. when expansion of mobile devices, and cloud computing and
viewing a website in a browser, determines whether it has making use of these to craft clever assaults. One of the most
been altered using computer vision algorithms. terrifying security risks facing the Internet is malware today.
It is changing and employing fresh strategies to attack
V. G. Davanzo , E. Medvet, A. Bartoli desktops and mobile devices. Additionally, the exponential
It is now a common issue for websites to be defaced. growth. The harm they do has grown in both volume and
Responses to these occurrences are frequently fairly sluggish complexity. These are capable of avoiding the earlier created
and occasionally prompted by user feedback because techniques for detection and mitigation that make it evident
corporations typically lack a thorough and constant that traditional cyber security must give way to cyber security
monitoring of the reliability of their websites. a more information. The goal of this study is to create a system for
methodical approach is undoubtedly a good idea. In this producing malware threat intelligence that can assess, an
regard, increasing availability is a tempting alternative and Early Warning System that can recognise and anticipate
services for performance monitoring with defacement malware attacks (EWS). Additionally, it displays the testing
detection. Motivated by these factors, in this study we of the proposed framework that is implemented by creating a
evaluate the effectiveness of various anomaly detection security-as-a-service prototype.
methods when faced with the issue of automatically spotting
web defacements. These strategies all create profiles. Y. Xiaozan Lyu1, Rodrigo Costas
automatically, depending on machine learning techniques, of Using the Big Data research field as a case study, we
the watched page and issue an alert when The content of the suggest a method for examining how academic subjects move
page does not match the profile. We evaluated their efficiency through interactions across audiences across various sources
with regard to false positives and On a dataset of 300 using altimetric. Altmetric.com and Web of Science provide
extremely dynamic web pages that we tracked for three the data used, with a concentrate on Twitter, Wikipedia, Blog,
months, false negatives were identified and a collection of News, and Policy. Author publication keywords. The
320 actual defacements keywords are taken as the primary issues of the publications
and the Altimetric hosts an online conversation about their
W. Youngho Cho audiences. Different methods are used to assess the
Techniques are becoming into more sophisticated, (dis)similarities between the subjects raised by the writers of
intelligent, and advanced. in the area of security studies, It is the publication and those viewed by online users. Results
a common and acceptable assumption in practise that indicate that there are significant differences overall
attackers are knowledgeable enough to find security flaws in connecting the two groups of Big Data research-related
security defence measures, preventing the identification of subjects. The primary deviation is Twitter, where tweets with
the defence systems and preventive measures. A series of frequent hashtags have a greater correlation with the
attacks known as "web defacement attacks" alter websites in keywords used by authors in publications. Blogs and News
an unauthorised manner. One of the serious continuous cyber are two online groups that provide a significant similarity in
risks that occur internationally is the use of web pages for the language utilised, while Wikipedia and Policy papers The
malevolent reasons. Such attacks can be detected using either largest differences across authors' approaches to and
server-based methods or client-based techniques. client- interpretations of big data research.
based strategies, each of which has advantages and
disadvantages. based on our thorough research on Using Z. Maria Ijaz Baig, Liyana Shuib* and Elaheh
current client-based protection techniques, we discovered a Yadegaridehkordi
serious security flaw that can be used to get access. by clever Big data is a crucial component of innovation that has
assailants. In this work, we outline the security flaw in the recently attracted significant attention both academics and
current client-based approaches that present unique practitioners' focus. Given the significance of the present
intelligent on-off web defacement and have a defined trend in the education sector is moving towards analysing the
monitoring cycle attacks that take advantage of this function of large info in this field. Numerous studies have
weakness. Next, we suggest use a random monitoring been done thus far to understand the use of big data in a
approach as develop two random monitoring defences as a variety of sectors and applications. However, an exhaustive
promising defence against such attacks. algorithms: (1) review of big data in education is still missing. Thus, the
Attack Damage-Based and (2) Uniform Random Monitoring objective of this study is to carry out a review of big data in
Algorithm (URMA) Automated Random Monitoring education to identify trends, group thematic areas of research,
(ADRMA). and draw attention to the shortcomings while offering
potential directions for the future. A systematic review
X. ekta gandotra, divya bansal, and sanjeev sofat process was used to examine 40 primary studies published
Network-capable ubiquitous computing devices have between 2014 and 2019 were made use of, and associated
evolved into the crucial cyber infrastructure for academics, data was gathered. The results indicated that there is an Over
government and business in daily life. The focus of the the past two years, there has been an upsurge in the number
cyberattacks against this vital infrastructure has switched to of research examining big data in education. The current
Political and commercial objectives are pursued, and this studies covered four primary study issues under large, it was

IJISRT22DEC1128 www.ijisrt.com 1203


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
discovered. Specifically, student behaviour and performance
data, modelling data, and instructional enhancement of the
educational system, big data integration, and data
warehousing the instruction. The majority of big data
research in education has been on students actions and
displays. The report also identifies research shortcomings and
depicts the directions for the future.

 Problem Statement
To develop a Malicious URL detecting system which
accurately detects and classifies the Benign and Malicious
URLs using Machine Learning and Deep Learning
Techniques.
Input: The dataset contains collection of malicious, benign,
spam, malware and defacement URLs in multiple formats like
csv, JSON, etc. Fig.1 Proposed Methodology
Output: Displays whether the URLs are Fraudulent and
legitimate based on features. III. MODULE DECOMPOSITION

II. PROPOSED METHDOLOGY

To accomplish this task, CNNs and RNNs have been


incorporated into neural network architectures. Here is the
diagram: Sequence generator architectures such as RNNs and
LSTMs can start by converting an image into a fixed-length
feature vector. This can be used to generate a set of words or
captions for an image. ResNet50 is the encoder we used for
this project. Millions of images in the ImageNet dataset were
classified into 1000 categories using pre-trained models. Its
weights are tuned to discriminate many things common in
nature, so to use this network effectively, remove the top layer
of 1000 neurons (for ImageNet classification) and replace it
with You can replace it with a linear layer containing the same
number of neurons as add . The number of neurons output by
the LSTM. An RNN consists of a series of Long Short-Term
Memory (LSTM) cells used to recursively generate captions
from input images. These cells use the concepts of repetition
and gates to remember information from past time steps. You
can watch or read this to know more. Finally, the encoder and Fig.2 Module Decomposition
decoder outputs are merged and passed to a dense layer and
finally to an output layer that predicts the next word based on  Modules include:
the image and current sequence.  Library Import and Dataset Collection: Dataset URL Use
.csv format for datasets. Then import various libraries
The proposed system is: needed by other modules. CSV files are converted to
 The first option is the graphical user interface (GUI). The pandas data frames. A URL is specified for feature
user intervenes in the system at this point. extraction.
 User must login or register when accessing here for the  Feature Extraction: IP Address – Provides the ability to
first time. create, manipulate and manipulate IPv4 and IPv6
 Users can then upload images and get descriptions. addresses and networks. re - A regular expression is used
 After the user enters a link or provides text, we use CNN to extract the features. Target audience: Create a simple
to extract features from the image and transform them into importable Python module that generates parsed WHOIS
fixed-length feature vectors. data for a given domain. Urllib: urllib is a package of
 After extraction, preprocess the image by changing its size, several modules for manipulating URLs. Urllib request is
orientation, color, brightness perspective. for opening and reading URLs.
 Feature Transformation: Feature values are assigned as 0
(legitimate) and 1 (malicious) based on conditions.
 Combine all features: Features extracted from different
sources are combined after the feature transformation step
for further processing.
 Split Feature Vector Data Set: Splits the data set into a
training data set and a test data set.

IJISRT22DEC1128 www.ijisrt.com 1204


Volume 7, Issue 11, December – 2022 International Journal of Innovative Science and Research Technology
ISSN No:-2456-2165
 Building Multiple Models: Building his 6 models that [6]. Zhiqiang Wang, Xiaorui Ren, Shuhao Li, Bingyan
incorporate both machine learning and deep learning Wang, Jianyi Zhang ,and Tao Yang “A Malicious URL
techniques. Detection Model Based on Convolutional Neural
 Scoring and comparing models: Models are scored using Network”
the Accuracy score metric or the Confusion metric. It is [7]. Cheng Cao and James Caverlee “Detecting Spam URLs
the ratio of the number of correct predictions to the total in Social Media via Behavioral Analysis”
number of input samples. Compare all models based on [8]. Christophe Chong, Daniel Liu, and Wonhong Lee
drawing and test accuracy. BLOCK DIAGRAM “Malicious URL Detection”
[9]. Xuan Dau Hoang and Ngoc Tuong Nguyen “Detecting
Website Defacements Based on Machine Learning
Techniques and Attack Signatures”
[10]. Ashit Kumar Dutta “Detecting phishing websites using
machine learning technique”
[11]. Frank Vanhoenshoven, Gonzalo Napoles, Rafael
Falcon, Koen Vanhoof and Mario Koppen “Detecting
Malicious URLs using Machine Learning Techniques”
[12]. Tiefeng Wu, Miao Wang , Yunfang Xi and Zhichao
Zhao “Malicious URL Detection Model Based on
Bidirectional Gated Recurrent Unit and Attention
Mechanism”
[13]. S. Markkandeyan, C. Anitha “Malicious URLs
detection system using enhanced Convolution neural
network”
[14]. Adebayo Oshingbesan Kagame Richard Aime
Munezero Courage O Ekoh “Detection of Malicious
Websites Using Machine Learning Techniques”
Fig.3 Block Diagram [15]. Immadisetti Naga Venkata Durga Naveen,
Manamohana K, Rohit Verma “Detection of Malicious
IV. CONCLUSION URLs using Machine Learning Techniques”
[16]. Malak aljabri, hanan S. Altamimi , shahd A.Albelali,
Malicious websites are a common social engineering maimunah al-harbi,haya T. Alhuraib, najd K. Alotaibi,
technique that mimics trusted URLs (Uniform Resource amal A. Alahmadi,fahd alhaidari “Detecting Malicious
Locators) and web pages. The goal of this project is to train URLs Using Machine Learning Techniques: Review
machine learning models and deep neural networks on the and Research Directions”
created data sets to predict malicious websites. It collects both [17]. Fuqiang Yu “Malicious URL Detection Algorithm
malicious and benign URLs of websites to form a dataset, based on BM Pattern Matching”
from which it extracts the desired URLs and content-based [18]. Kevin Borgolte, Christopher Kruegel, Giovanni Vigna
functionality of the website. Measure and compare the “Meerkat: Detecting Website Defacements through
performance level of each model. This project aims to use Image-based Object Recognition”
machine learning and deep learning techniques to better [19]. Xuan Dau Hoang and Ngoc Tuong Nguyen
predict malicious URLs. “DetectingWebsite Defacements Based on Machine
Learning Techniques and Attack Signatures”
REFERENCES [20]. Trong Hung Nguyen1,Xuan Dau Hoang2,Duc Dung
Nguyen “Detecting Website Defacement Attacks using
[1]. Clayton Johnson, Bishal Khadka, Ram B. Basnet Web-page Text and Image Features”
”Towards Detecting and Classifying Malicious URLs [21]. Kevin Borgolte “Identifying and Preventing Large-scale
Using Deep Learning” Internet Abuse”
[2]. Vinayakumar R, Sriram S, Soman KP, and Mamoun [22]. G. Davanzo , E. Medvet, A. Bartoli “Anomaly detection
Alazab ”Malicious URL Detection using Deep techniques for a web defacement monitoring service”
Learning” [23]. Youngho Cho “Intelligent On-O_Web Defacement
[3]. Shantanu, Janet B, Joshua Arul Kumar R “Malicious Attacks and Random Monitoring-Based Detection
URL Detection: A Comparative Study” Algorithms”
[4]. Mohammad SaifulIslam, Mamun, Mohammad Ahmad [24]. ekta gandotra, divya bansal, and sanjeev sofat “A
Rathore, Arash Habibi Lashkari, Natalia Stakhanova, framework for generating malware threat intelligence”
and Ali A. Ghorbani “Detecting Malicious URLs Using [25]. Xiaozan Lyu1, Rodrigo Costas “How do academic
Lexical Analysis” topics shift across altmetric sources? A case study of the
[5]. Tie Li a, Gang Kou b, Yi Peng a “Improving malicious research area of Big Data”
URLs detection via feature engineering: Linear and [26]. Maria Ijaz Baig, Liyana Shuib and Elaheh
nonlinear space transformation methods” Yadegaridehkordi “Big data in education: a state of the
art, limitations, and future research directions”

IJISRT22DEC1128 www.ijisrt.com 1205

You might also like