Sans Detecting Application Layer Ddos Attacks Using Tls Fingerprinting
Sans Detecting Application Layer Ddos Attacks Using Tls Fingerprinting
gh
Detecting Application Layer DDoS Attacks Using
Ri
TLS Fingerprinting
ll
Fu
ns
ai
GIAC (GCIA) Gold Certification
et
Author: Alejandro Aucestovar, [email protected]
rR
Advisor: Domenica Crognale
ho
ut
Accepted: 12/10/2021
,A
te
Abstract
itu
Application layer DDoS attacks are some of the most complex and devastating attacks on
st
their advantage so that identification and mitigation do not happen easily. Previous
NS
research has had different levels of success at identifying DDoS attacks and
differentiating them from legitimate human traffic as well as legitimate flash flood
SA
methodology and JA3 fingerprinting technique were tested against the Canadian Institute
of Cybersecurity’s DDoS dataset created in 2019. TLS Fingerprinting successfully
22
gh
Ri
1. Introduction
ll
Fu
The rise in consumer consumption and e-commerce via websites and applications
has risen significantly in the last two decades and is a major revenue focus for many
ns
companies. Some companies like Amazon exist almost entirely online and rely on their
ai
systems and applications to be always available to the public. For this reason, distributed
et
denial of service attacks have been a major concern for security experts, business leaders,
rR
and e-commerce and financial institutions worldwide for the last two decades. The
ho
financial impact of a DDoS attack on an e-commerce or financial website can range
ut
vastly depending on the duration of the attack and the assets of the victim company. To
,A
protect their online presence, companies are forced to decide how to mitigate DDoS
te
attacks with the assistance of cyber security providers that provide protection or with
itu
cloud service providers that provide scalability, load balancing, and global traffic
st
management.
In
Companies such as Akamai, Cloudflare, Radware, and NexusGuard are just a few
NS
cyber security providers that companies call upon for DDoS protection. These
SA
companies provide a variety of services that can identify and mitigate DDoS attacks.
These services include serving a company’s static content on the service provider's
e
Th
network, detecting high traffic rates or behavioral patterns via complex algorithms, or
22
using proxies to inspect and detect abnormalities before the traffic reaches the real e-
commerce website. Since DDoS protection is highly complex and does not always work,
20
especially with the rise of amplification via botnets, companies have increasingly chosen
©
to seek help from cloud providers such as Google and Amazon. Cloud service providers
are often a good choice for companies since the cloud provider is usually responsible for
identifying and mitigating DDoS attacks on their online systems. Cloud providers can
also spin up new servers and provide load balancing to those new servers with almost no
delay or impact to the company website. Even though cloud providers can provide
availability services, they are not exempt from the effects of a DDoS attack, especially
those initiated by botnets such as Mirai botnet. The best approach to combatting DDoS
attacks is to take a hybrid approach where the company hires cyber security and cloud
availability services for their e-commerce sites and applications.
gh
Ri
DDoS attacks come in many different shapes and sizes, and their successful
ll
detection and mitigation must match the shape and size of those attacks. In other words,
Fu
if a DDoS attack utilizes an open stack interconnection (OSI) layer four, then the
ns
detection and mitigation must occur within OSI layer four. Cyber security providers are
ai
particularly adept at detecting and mitigating layer four and layer five DDoS attacks.
et
Cyber security providers attempt to profile traffic patterns that match certain typical
rR
human traits and behaviors and compare those behaviors to the behaviors of bots. DDoS
ho
attacks at the application layer or the OSI layer seven present a more complex problem
ut
for cyber security companies because of the additional packet inspection needed to view
,A
the data and establish normal behaviors of legitimate customers. The use of encryption
via transport layer security (TLS) or secure sockets layer (SSL) makes this problem even
te
itu
more complex.
st
Police agencies across the globe have used fingerprinting as part of their
In
characteristics individual humans possess. This research will explore using various
SA
aspects of the TLS handshake to fingerprint the traffic from legitimate and illegitimate
users to DDoS attacks such as HTTP Get Flood over SSL. When used in conjunction
e
communications with their customers as early as 1999. However, TLS encryption was
not widely adopted until much later. The purpose of TLS was to provide an upgrade to
its predecessor, SSL 3.0, which was susceptible to POODLE attack, by using a
standardization protocol that can be used to define cipher suites with symmetric
cryptography, which include RC4, Triple DES, and AES (RFC 2246). Several
subsequent versions of TLS (TLS1.1, TLS 1.2, TLS1.3) have been developed and
superseded for various reasons and shortcomings, but the purpose of TLS, to provide
communications privacy and prevent eavesdropping, has remained the same. RFC 2246
outlines the process undertaken by both client and server to establish secure
gh
Ri
communication with encryption, authentication, and integrity through a handshake and
ll
record process. The primary function of the recording process is to encrypt and send the
Fu
data using the information from the handshake process. The TLS handshake process
ns
begins after the transport connection protocol (TCP) handshake and is always transmitted
ai
in the clear with little disruption to the services. TLS fingerprinting takes advantage of
et
the data transmitted in the clear during the handshake process to obtain details about the
rR
client and the server and preserve the messages' privacy.
ho
1.2. Research Goals
ut
Researchers and security professionals attempt to differentiate legitimate from
,A
illegitimate traffic in DDoS attacks by identifying normal user behavior and setting it as
te
the baseline. Once this baseline is established, all traffic that does not fit the baseline is
itu
creates a problem for researchers since the traffic is legitimate but behaves like
In
illegitimate bot traffic. Detection and mitigation of DDoS attacks have focused on the
NS
characteristics of either user attributes, traffic attributes, or traffic rates (Karanpreet et al.,
SA
2016). A combination of user attributes such as user-agents, header data, the number of
pages visited, sequence of pages viewed, or rate of received/sent packets have been
e
Th
Relationships between humans have always been based on how well we know
©
each other and trust each other. Generally speaking, the more we know about each other,
the more trust we provide. This concept derived from the human trust should be applied
to our internet interactions to measure client/server trust. For example, in a banking
website, a user who provides personal details and creates an account is provided more
access to the website than a customer who does not. These customers with accounts are
trusted more because the banking website knows more about them. The user even
reciprocates this trust to the bank since the bank is holding their money. Researchers and
security professionals apply this concept to incoming traffic as they focus on the user and
traffic characteristics that make up the ‘recipe’ identifying legitimate, trusted traffic. This
gh
Ri
concept has shown varying results in identifying and mitigating application-layer DDoS
ll
attacks (Park et al., 2021, Zhao et al., 2018, Wei et al., 2020, Suchacka et al., 2020). Still,
Fu
TLS fingerprinting could add trust and fidelity to this process by providing more
ns
information about the user initiating the traffic. The more we (website owners, security
ai
professionals, business leaders, etc.) know who is coming to our websites and what they
et
are doing, the better we will get to know them (users, clients, etc.) and attach the
rR
appropriate level of trust.
ho
2. DDoS Identification Recipe
ut
,A
Application layer DDoS attacks can generally be categorized into high-rate and
te
low-rate attacks (Mirkovic et al., 2004). Low-rate attacks such as Slowloris can begin at
itu
slow or constant rates but do not stop, which bottles up server resources. High-rate
st
attacks such as HTTP Get Floods can vary with elevated rates but still impact the server’s
In
attacks are far more complex than lower layer attacks because of all the valid components
SA
that the traffic contains, such as valid TCP connections, valid IPs, and valid requests.
They do this under the protection of TLS encryption. Researchers and security
e
Th
professionals have sought help from web application firewalls (WAF) to try and detect
and mitigate application-layer DDoS attacks. WAFs are incredibly helpful in identifying
22
layer seven attacks and are considered “must-haves” when building security architecture
20
around web applications— but they cannot catch everything. Even though most WAFs
©
are well suited for signature detection, they cannot detect attacks based on behavioral
analysis, which is needed in DDoS attacks. Researchers must rely on identifying user
and traffic characteristics attributable to normal human traffic and classify other traffic as
bot traffic.
gh
Ri
a website, no matter the purpose of the website, should be different from the interaction
ll
and traffic from bot-related traffic.
Fu
2.1.1. Human Versus Bot Traffic
ns
According to Suchaka et al. (2020), technological advances have made bot traffic
ai
extremely difficult to distinguish from legitimate human web traffic. The bots, botnets,
et
and shilling attacks try to mimic human traffic. Bots are similar to scripts in that they
rR
help humans with repetitive tasks that may not be very complicated but are high in the
ho
number of events to be completed. According to Lagopulous et al. (2017), bot traffic was
ut
identified by a high number of repeated requests, long session durations, and a high
,A
number of total requests. Similarly, Radware Bot Manager (n.d.) discusses how bot
te
traffic is detected when there are many hits from a single IP during a short time. Human
itu
traffic would have a significantly lower number of repeated requests, uneven session
st
durations, and fewer total requests. Further, Sreeram et al. (2019) identified that bot
In
traffic.
SA
Human traffic resembles human behavior in that the session durations change
from website to website, dictated by personal interest. Humans tend to request only
e
Th
interesting pages, whereas bots request all web pages or the same pages across different
22
application layer 7 DDoS attacks. They provided the following characteristics pertaining
to human and bot traffic: “Some actions or features usually associated with the legitimate
users are as follows:
gh
Ri
• likely to request for popular web pages;
ll
• hardly ever repeat their web access patterns;
Fu
• rely on legitimate web browsers to connect to the server;
ns
ai
• exhibit widely dispersed geographical distribution.
et
Some of the few properties related to attacking bots are as follows:
rR
• infected by the presence of malware;
ho
•
ut
await attacker’s command before initiating an attack;
,A
• provide false identity by faking user-agent strings;
te
• almost similar access patterns among all attacking bots;
itu
• access behavior more often than not deviates from legitimate users;
st
In
Flash flood events are those events that are generated by a high number of
22
requests from legitimate users, usually as a result of a marketing campaign, new web
20
page or product, or new services but behave much like illegitimate DDoS traffic. In a
©
sense, flash flood events mimic not only human behavior but also bot frequency and
velocity. According to Khalaf et al. (2021), flash flood events can have the same impact
as illegitimate DDoS traffic, overloading the server, computational performance, and
exhausting resources. However, it is created by legitimate users trying to access the
website. According to Singh et al. (2018), bot-generated DDoS traffic has almost
identical inter-arrival time rates as flash flood traffic. Still, it is different in that the
number of instances per IP address is much larger in DDoS attacks. In other words,
although requests are coming in almost simultaneously in both DDoS and flash flood
events, there are more instances, connections, or sessions per IP during a DDoS attack.
gh
Ri
2.1.3. Selected Research to Emulate
ll
Throughout the research process of this whitepaper, the work of Karanpreet et al.
Fu
(2017) was a highly relevant contribution to the research and body of work on
ns
application-layer DDoS attacks. Karanpreet and associates conducted a wide-ranging
ai
survey of all available application layer DDoS research from 2005 to 2015, focusing on
et
flood attacks. After reviewing all the selected studies within their work, only the newest
rR
studies conducted in 2015 were selected for review. Out of the five selected studies, only
ho
two studies used multiple detection attributes that had either a high likelihood of
ut
detection for typical human values or typical bot values based on Karanpreets’ analysis.
,A
Saleh et al. (2015) used a combination of the total number of requests and the total
number of hot pages and calculated a ratio which was then used to multiply by the log of
te
itu
the URI count and total count ratio. Liao et al. (2015) used a combination of all users'
total number of requests and the number of total objects. They created a ratio that was
st
In
then applied to the likelihood of switching between web pages. Since the dataset used for
NS
this research does not seem to use hot pages, this whitepaper will use the methodology
described by Liao.
SA
Conducting research in the analysis of cybers attacks has always been difficult
because it is highly unlikely that a company allows its data to be shared with the public
22
and be seen by potential malicious persons. In the past, DDoS researchers have taken
20
one of two approaches, creating their data in their control lab environments or obtaining
©
publicly accessible data sets. The obvious problem with researchers creating their data in
a controlled lab is that there is little trust the data resembles actual DDoS data in the wild.
A tool such as hping3 from Kali Linux is one of the tools used by actors to generate
DDoS attacks. Hping3 tools allow for a script to be developed and executed to generate
hundreds of thousands of HTTP GET and POST requests to single or multiple targets.
Another tool used by both professionals and bad actors is the HTTP Unbearable Load
King (HULK) created by Barry Schteiman, which was designed to generate and send
volumes of obfuscated traffic to overwhelm the web servers and act as a stress tester.
Although the tools used in the lab environment are likely the same as those used by
gh
Ri
attackers, professionals still distrust professionals that the lab traffic is the same as DDoS
ll
attack traffic.
Fu
Publicly accessible data composed of DDoS attack traffic does not carry the same
ns
distrust from researchers and security professionals. The ability to analyze and take apart
ai
real DDoS traffic is a great advantage to researchers and professionals because they can
et
look deep into the attack's process, generation, and logic. This deep insight into DDoS
rR
traffic can provide the sought-after answers that the community desire for the
ho
identification and mitigation of DDoS attacks, theoretically. Unfortunately, all the
ut
publicly accessible datasets are rather old, dating back to 1998 when the internet was just
,A
getting started. Sreeram et al. (2017) use a dataset created in 2007 by the Center for
te
Applied Internet Data Analysis (CAIDA) from the University of California's San Diego
itu
attacks that attempted to discriminate a DDoS attack from a flash flood event. Behal and
NS
associates found that all 16 studies used the 2007 CAIDA dataset or the 1998 FIFA
SA
World Cup dataset, which the Network Research Group maintained at Lawrence
Berkeley National Laboratory, Berkley, CA. The internet has certainly changed since
e
1998 and 2007, from the increased use of internet services to more secure protocols such
Th
as TLS encryption. TLS 1.2 was released to the public in August 2008 and was not
22
widely used until recent years. There is no question that researchers and security
20
DDoS attacks.
gh
Ri
2.1.5. Canadian Institute for Cybersecurity (CIC) DDoS Dataset 2019
ll
Fu
This research will utilize a newer dataset for experimentation, using some of the
characteristics already found in previous application layer DDoS studies. It will also
ns
apply TLS fingerprinting techniques to increase fidelity in identifying these types of
ai
attacks. The CICDDoS2019 dataset was created by researchers from the University of
et
New Brunswick after reviewing publicly available DDoS datasets since 2007 and finding
rR
numerous shortcomings. Researchers then used the CICFlowMeter-V3 tool to address
ho
many of the shortcomings of previous datasets to provide a more realistic and
ut
representative DDoS dataset of today’s internet. The CICDDoS2019 dataset proposes a
,A
new detection and family classification approach for application-layer DDoS attacks.
te
The dataset was created in two days which was comprised of a training day and a testing
itu
day. During the training day, traffic was generated for 12 DDoS attacks, including NTP,
st
DNS, LDAP, MSSQL, NetBIOS, SNMP, SSDP, UDP, UDP-Lag, WebDDoS, SYN, and
In
TFTP. During the testing day, seven attacks were generated, including PortScan,
NS
NetBIOS, LDAP, MSSQL, UDP, UDP-Lag, and SYN. The researchers created an
SA
implementation of TLS 1.2. According to RFC 5246, TLS 1.2 has two parts, the
handshake protocol, and the record protocol. The handshake protocol is a process where
the client and server communicate to establish an encrypted channel by exchanging and
negotiating their allowable cipher mechanisms and keys. The handshake process takes
place in plaintext, is visible to anyone sniffing out the traffic, and is valuable to the TLS
fingerprinting process. Gancheva et al. (2020) provide an excellent illustration of the TLS
handshake process in Figure 2 below.
gh
Ri
ll
Fu
ns
ai
et
rR
ho
ut
,A
te
Gancheva et al (2020) Figure 2, p. 17
itu
The exchange of hello messages and corresponding data is sent in plain text along
st
In
with the server ChangeCipherSpec. Researchers can use these parameters to conduct
NS
TLS fingerprinting of both the client and the server, which can be applied to future
connections and subsequent sessions.
SA
e
fingerprint the clients and servers that are making connection requests to each other. The
20
way a client makes a connection and executes the handshake process should be the same
©
when making future or subsequent connection requests to the same server. Similarly,
servers can be expected to also behave in the same manner when responding to future or
subsequent requests from a previous client. This server behavior can be seen in the TLS
fingerprinting technique known as Markov Chain fingerprinting. The Markov technique
is complex but focuses on the server responses based on the message type sequence and
applies statistical analysis to either first-order or second-order chains.
gh
Ri
applied to either a host-based, which uses information from HTTP header once the server
ll
decrypts the connection, or flow-based, which uses cipher suites from HTTP connections
Fu
and combines the user-agent from the HTTP connection.
ns
A third TLS fingerprinting method is the JA3/JA3S method developed by
ai
Salesforce engineers John Althouse, Jeff Atkinson, and Josh Atkins. This method
et
converts fields from the ClientHello message such as version, accepted ciphers, and other
rR
fields into decimals, and then MD5 hashed as the first step of the fingerprinting process.
ho
Secondly, the same process is conducted on the server response using the data from the
ut
ServerHello message (Althouse, 2019). Security professionals have received the
,A
JA3/JA3S fingerprinting technique with open arms, especially those researching malware
te
analysis. Researchers can fingerprint the connection behaviors of well-known hacking
itu
tools such as Kali Linux and Metasploit and some indicators of compromise from
st
malware embedded within the software. Unfortunately, bad actors are now fully aware
In
of this TLS fingerprinting technique. They have found ways around the process by
NS
to evade specific malware signature detection. The goal is to determine if the client is a
Th
legitimate user such as a human or an illegitimate user such as a bot. In other words,
22
even if the JA3/JA3S fingerprint does not match the signature of a well-known hacking
20
tool or malware, the JA3/JA3S fingerprint should remain the same with multiple,
©
subsequent, or future connections from the same client. Adding the details of the
JA3/JA3S fingerprinting process as a property or characteristic of an illegitimate or
legitimate user may provide fidelity to application layer DDoS detection.
4. Analysis
The CICDDoS2019 dataset was analyzed with Wireshark and Tcpdump. The
detection attributes section focuses on the encrypted data extracted from the WebDDoS
data using Tcpdump. TLS fingerprinting was conducted using a Python script that used
gh
Ri
the JA3/JA3S fingerprinting technique and focused on the TLS handshake data extracted
ll
from the WebDDoS traffic.
Fu
4.1. WebDDoS Data Preparation
ns
The entire CICDDoS2019 dataset is comprised of two days of DDoS attack traffic
ai
split into approximately 146 large PCAP files, which are timestamped directly on the
et
rR
CIC’s dataset webpage. The WebDDoS was conducted from 13:18 to 13:29 on March
11, 2018, and was documented in a PCAP file named SAT-03-11-2018_0144. The
ho
victim network was 192.168.50.0/24 with only 192.168.50.6 and 192.168.50.8-9 active
ut
during the WebDDoS attack. The following illustration captures the web architecture of
,A
the victim network and the traffic flow of the DDoS attacks directly taken from the
te
CICDDoS2019 website.
itu
st
In
NS
SA
e
Th
22
20
©
The architecture design used by the CIC did not utilize a firewall, intrusion
detection or intrusion prevention system, or any other modern security controls. There is
an opportunity to use the F5-Modem as a firewall or filtering protocol that could have
been utilized to inspect the traffic. Still, the researchers did not choose to do this without
gh
Ri
explanation or comment. The CIC researchers did not provide a source internet protocol
ll
address or domain for the third party company, nor did they indicate legitimate users.
Fu
4.2. Detection Attributes
ns
According to Liao et al. (2015), their detection algorithm was derived from two
ai
derived attributes: the request interval sequence and the request frequency sequence. In
et
rR
order words, the distinction between DDoS illegitimate traffic and legitimate human
traffic depends on the average popularity of web objects and the average transition
ho
between each object. The average popularity of web objects is the ratio of all users' total
ut
number of requests over the number of total objects. The average transition between
,A
each object is the time between the last request and the following request and the
te
likelihood of visiting or requesting those more popular web objects. Liao et al. (2015)
itu
rationalize that illegitimate DDoS traffic is generated by bots and will have a regular or
st
rhythmic pattern, whereas legitimate human traffic is highly irregular and unpredictable.
In
Furthermore, legitimate human traffic resembles human behavior in that only interesting
NS
web objects will be requested. In contrast, illegitimate bot traffic will request all web
SA
objects or a consistent number or pattern of objects, but consistently more than humans.
e
The request frequency sequence has always been a pivotal problem for
22
researchers and security professionals in the HTTP Get Flood over SSL/TLS. Encryption
20
hides the fine details of a client's request, and this shortcoming continued in this research.
Without the SSL/TLS certificates used during the testing and creation of the
©
CICDDoS2019 dataset, the HTTP requests to the victim network could not be inspected.
The dataset creators were contacted to retrieve TLS certificates via personal and work
email but did not respond. Even though the fine details of the requests were not
available, those connections and requests to the victim network were still analyzed.
Since this characteristic focuses on the request frequency sequence, all traffic during the
TCP and TLS handshake must be excluded. Tcpdump was used to look at the traffic over
port 443 to the victim network and isolated the encrypted application data by filtering the
packets with a 17 in the first nibble of the TCP header. From 15:37:01UTC to
17:15:12UTC, there were 30274 requests to the victim network, from approximately 620
gh
Ri
distinct source IPs, at a rate of approximately 308.91 requests per minute. During the
ll
same time frame, there were 18846 encrypted requests to 192.168.50.9, 4786 encrypted
Fu
requests to host 192.168.50.8, 6340 requests to host 192.168.50.6, and 302 requests to
ns
192.168.50.4.
ai
4.2.2. Request Interval Sequence (RIS)
et
The request interval sequence was easier to determine since there was no need to
rR
decrypt any traffic, and the focus was on the timestamps of each request. Using
ho
Tcpdump, the data were extracted and transferred into a Microsoft Excel spreadsheet
ut
with time stamps and time deltas in microsecond precision (Tcpdump -ttt) into four
,A
separate files. The largest file contained all the requests sent to the victim network
te
192.168.50.0/24, and the other files contained the traffic sent to the other three victim
itu
legitimate and illegitimate traffic is that greater request intervals characterize legitimate
In
traffic. Table 1 depicts statics based on the time interval between the current request and
NS
There were over 30,000 requests in all four files, and after visually inspecting all
©
the requests, approximately 65% of the requests were sent to host 192.168.50.9. Even
though most of the traffic was sent to host 50.9, three timeframes in each of the hosts
(50.8 and 50.6) were compared to understand the request interval sequence. Table 2
depicts the one-minute time frames where the number of requests was higher in all three
victim hosts. Compared to the overall victim network statistics, the average request
interval sequence for both hosts, 50.8 and 50.6, was significantly smaller. Host 50.6 had
a significantly faster request interval sequence during the timeframe 17:04 UTC. There
were only five source IPs, 172.217.10.226, 172.217.10.2, 172.217.9.226, 172.217.11.2,
gh
Ri
216.58.219.226, sending traffic during this time. These source IPs are good candidates to
ll
be classified as illegitimate.
Fu
.50.9 15:42 15:58 16:03
ns
Average 0.425738 0.17495 0.575747
Std Dev 3.838843 2.011933 3.596295
ai
et
.50.8 16:10 16:24 17:25
rR
Average 0.012856 0.008239 0.128753
Std Dev 0.092507 0.058321 1.347081
ho
.50.6 17:04 17:09 17:20
ut
Average 0.000004 0.00019 0.002047
,A
Std Dev 0.000001 0.000085 0.000641
Table 2. Request Interval Sequence for Select One-Minute Timeframes
te
itu
legitimate users. Some possible legitimate users were discovered searching the dataset
In
via Tcpdump, focusing on single-source addresses with no more than 50 requests. Table
NS
3 depicts request interval sequence statistics for possible legitimate users. Host 50.9 had
SA
50 requests from host 13.35.78.80 during the entire duration of the WebDDoS attack,
e
with an average of approximately 11.79 seconds between each request. Compared to the
Th
average request interval sequence of host 13.35.78.80, this request interval sequence is
22
gh
Ri
4.3. JA3 Fingerprinting
ll
Fu
The JA3/JA3S fingerprinting technique profiles both candidates' legitimate and
illegitimate users from the CICDDoS2019 dataset. To calculate JA3/JA3S fingerprints, a
ns
Lua plugin for Wireshark was used. I then created two filters, "JA3" and "JA3S," which
ai
et
were then used to display as columns for easy tracking. The illegitimate candidate hosts
rR
sending traffic to host 50.6 were isolated and then exported into a smaller PCAP file to be
examined with Tcpdump. The JA3/JA3S calculations for the candidate illegitimate hosts
ho
showed three distinct JA3/JA3S fingerprints that all the candidate hosts used. Table 4
ut
depicts the details on the JA3 calculations and the server the traffic was sent to per the
illegitimate candidate hosts.
,A
te
itu
334da95730484a993c6063e36bc90a47 192.168.50.8
NS
eca9b8f0f3eae50309eaf901cb822d9b 192.168.50.6
SA
334da95730484a993c6063e36bc90a47 192.168.50.8
Th
eca9b8f0f3eae50309eaf901cb822d9b 192.168.50.6
22
Isolating the JA3 fingerprints used by all illegitimate candidate hosts as a filter in
Wireshark showed the same calculation was made in other hosts' connections to the
victim network, as shown in Image 1 below. The JA3 Fingerprint
“eca9b8f0f3eae50309eaf901cb822d9b” was calculated in 540 different connections from
67 clients to the victim network. The other JA3 Fingerprint calculations observed from
gh
Ri
the candidate illegitimate hosts were seen across thousands of connections and from over
ll
100 client hosts. Further research was conducted by searching the online JA3 Fingerprint
Fu
database accessible at ja3er.com found that JA3 fingerprint
ns
“b20b44b18b853ef29ab773e921b03422” showed the fingerprint matched fingerprints of
ai
the onion router (TOR) and multiple operating systems such as Ubuntu, Windows NT,
et
and macOS.
rR
ho
ut
,A
te
itu
st
In
NS
SA
e
Th
22
20
©
Isolating the traffic from the candidate legitimate host 13.35.78.80 in Wireshark
showed the JA3 fingerprint for this host was “76cc3e2d3028143b23ec18e27dbd7ca9”.
Filtering by this JA3 Fingerprint in Wireshark revealed the same calculation was seen in
gh
Ri
245 different encrypted connections originating from 26 different client hosts. Again, the
ll
online JA3 Fingerprint database at ja3er.com was searched for this fingerprint, but there
Fu
was no information found. Image 2 depicts the details of this JA3 fingerprint shared by
ns
the 26 different hosts.
ai
et
rR
ho
ut
,A
te
itu
st
In
NS
SA
e
Th
22
gh
Ri
internet protocol addresses tends to raise concerns. Previous research on layer seven
ll
DDoS attacks has also shown the lack of entropy in IP locations, a strong indicator of
Fu
illegitimate DDoS traffic (Zhao et al., 2016, Swaminathan et al., 2014, Tongguang et al.,
ns
2013, Siracuscano et al., 2021). Utilizing the information and data used during the TLS
ai
handshake may have some ethical and contractual implications for security practitioners.
et
Still, a greater concern would be the storage and access to that data for future use. It is
rR
advisable for security practitioners and business leaders to disclose data collection during
ho
all internet communications and gain consent from their users before actively collecting
ut
data.
,A
Previous layer seven DDoS research has shown that attackers are imitating human
te
behavior to bypass security and corporate security, and the same trend is happening with
itu
TLS Fingerprinting. Attackers are masking their TLS fingerprints to avoid detection as
st
they attempt to avoid detection and gain access into networks. In the arena of DDoS
In
attacks, attackers changing or masking the TLS Fingerprint may make it easier to identify
NS
illegitimate traffic. Trust is at the center of the TLS Fingerprinting technique, where a
SA
client and server will always respond and communicate the same way, and if that
changes, then trust is lost. When trust is lost, relationships are broken, contracts are
e
This research paper analyzed two known identifying characteristics of layer seven
DDoS attacks and added TLS fingerprinting as a third characteristic. Still, it was not
©
identical to previous research. Liao et al. (2015) used machine learning (ML) and
artificial intelligence (AI) to calculate RIS and RFS using an algorithm they called a
sparse vector decomposition and rhythm matching (SVD-RM). Applying the same
technique in future research that Liao et al. used would be more effective in
differentiating illegitimate DDoS traffic from legitimate flash flood traffic. To do this,
future researchers would first need to find suitable legitimate and illegitimate traffic so
the IA/ML can establish a baseline.
Future research should focus on looking at real-world data where TLS certificates
are available so that the content of each request can be analyzed and taken into
gh
Ri
consideration. Applying TLS Fingerprinting to DDoS analysis is part of the reasoning
ll
for providing more information about who the users and clients requesting web pages are.
Fu
If researchers can see the content of each request, they can make better-informed
ns
decisions about legitimate or illegitimate traffic. Previous layer seven DDoS attack
ai
research focused on web page details such as a user requesting the robots.txt file, passing
et
CAPTCHA challenges, hot pages, out of date pages, hyperlink depth, etc. (J. Wang,
rR
2011, Behal et al., 2016, Park et al., 2021, K. Singh, 2016).
ho
Future research could also study the applicability of TLS Fingerprinting with
ut
other identifying layers seven DDoS characteristics. Characteristics such as passing
,A
CAPTCHA challenges and access to hot pages could be combined with TLS
te
Fingerprinting to increase fidelity in differentiating illegitimate from legitimate traffic.
itu
6. Conclusion
SA
TLS Fingerprinting such as JA3 Fingerprinting can provide valuable insight about
the users requesting web pages and web objects which can help security practitioners in
e
Th
their decisions about legitimate and illegitimate traffic. Like most applications in cyber
security, the detection and mitigation of DDoS attacks must have multiple filters and
22
multiple tests to ensure high fidelity of results. TLS Fingerprinting helps security
20
practitioners identify layer seven DDoS attacks, but future research should also include
©
mitigation techniques.
gh
Ri
References
ll
Fu
Althouse, John. (2019). TLS Fingerprinting with JA3 and JA3S. Medium. Retrieved
ns
ai
with-ja3-and-ja3s-247362855967
et
Behal, Sunny, Kumar, Krishan, & Sachdeva, Monika. (2016). Discriminating Flash
rR
Events from DDoS Attacks: A Comprehensive Review. International Journal of
ho
ut
Network Security, Vol 19, No 5, 2017, pp 734-741.
,A
https://doi.org/10.6633/IJNS.201709.19(5).11
te
Defending HTTP web servers against DDoS attacks through busy period-based attack
itu
st
8(7). https://doi.org/10.3837/tiis.2014.07.018
NS
Gancheva, Z., Sattler, P., & Wustrich, L. (2020). TLS Fingerprinting Techniques.
SA
https://doi.org/10.2313/NET-2020-04-1_04
22
How to detect bot traffic and identify bot threats. Radware Bot Manager. (n.d.). Retrieved
20
bot-traffic/.
Iman Sharafaldin, Arash Habibi Lashkari, Saqib Hakak, and Ali A. Ghorbani. (2019)
gh
Ri
Jaafar, G. A., Abdullah, S. M., & Ismail, S. (2019). Review of recent detection methods
ll
for an HTTP DDOS attack. Journal of Computer Networks and Communications,
Fu
2019, 1–10. https://doi.org/10.1155/2019/1283472
ns
ai
J. Wang, X. Yang, and K. Long, "Web DDoS detection schemes based on measuring
et
user's access behavior with large deviation," In Proc. IEEE Global Telecommun.
rR
Conf., 2011, pp. 1-5
ho
ut
Karanpreet, Singh., Paramavir, Singh., and Kumar, Krishan S. (2017). Application Layer
,A
HTTP-Get Flood DDoS Attacks: Research landscape and challenges. Computers
te
and Security. https://doi.org/10.1016/j.cose.2016.10.005
itu
denial of service attacks,” Comput. Security, vol. 56, pp. 111-139, 2016
NS
Liao, Qin., Li, Hong., Kang, Songlin., and Liu, Chuchu. (2015). Application layer DDoS
SA
attack detection using a cluster with a label based on sparse vector decomposition
e
Th
3120. https://doi.org/10.1002/sec.1236
22
20
Park, S., Kim, Y., Choi, H., Kyung, Y., & Park, J. (2021). HTTP DDoS flooding attack
©
https://doi.org/10.1587/transinf.2021edl8022
Prasad, K. M., Reddy, A. R., & Rao, K. V. (2016). Anomaly-based real-time prevention
https://doi.org/10.17485/ijst/2016/v9i27/87872
gh
Ri
Rahman, O., Quraishi, M. A., & Lung, C.-H. (2019). DDoS attacks detection and
ll
mitigation in SDN using machine learning. 2019 IEEE World Congress on
Fu
Services (SERVICES). https://doi.org/10.1109/services.2019.00051
ns
ai
Ramezani, A., Khajehpour, A., & Siavoshani, M. J. (2020). On multi-session website
et
fingerprinting over TLS handshake. 2020 10th International Symposium
rR
OnTelecommunications (IST). https://doi.org/10.1109/ist50524.2020.9345817
ho
ut
Saleh, M. A., & Abdul Manaf, A. (2015). A novel protective framework for defeating
,A
HTTP-based denial of service and distributed denial of service attacks. The
te
Scientific World Journal, 2015, 1–19. https://doi.org/10.1155/2015/238230
itu
Singh, K. J., Thongam, K., & De, T. (2018). Detection and differentiation of application-
st
In
layer DDoS attack from Flash events using fuzzy‐ga computation. IET
NS
Singh, K., Singh, P., & Kumar, K. (2017). Application layer HTTP-get floods DDoS
e
Th
attacks: Research landscape and challenges. Computers & Security, 65, 344–372.
https://doi.org/10.1016/j.cose.2016.10.005
22
20
Sreeram, I., & Vuppala, V. P. (2019). HTTP flood attack detection in application layer
©
https://doi.org/10.1016/j.aci.2017.10.003
Suchacka, G., & Iwański, J. (2020). Identifying legitimate web users and bots with
gh
Ri
Wei, F., & Yuqin, W. (2020). DDoS attack real-time defense mechanism using Deep Q-
ll
Learning Network. International Journal of Performability Engineering, 16(9),
Fu
1362. https://doi.org/10.23940/ijpe.20.09.p5.13621373
ns
ai
Zhao, Y., Zhang, W., Feng, Y., & Yu, B. (2018). A classification detection algorithm
et
based on joint entropy vector against Application-layer DDoS attack. Security and
rR
Communication Networks, 2018, 1–8. https://doi.org/10.1155/2018/9463653
ho
ut
,A
te
itu
st
In
NS
SA
e
Th
22
20
©